13731 1727203821.28174: starting run ansible-playbook [core 2.17.4] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-bGV executable location = /usr/local/bin/ansible-playbook python version = 3.12.5 (main, Aug 23 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-1)] (/usr/bin/python3.12) jinja version = 3.1.4 libyaml = True No config file found; using defaults 13731 1727203821.28587: Added group all to inventory 13731 1727203821.28589: Added group ungrouped to inventory 13731 1727203821.28593: Group all now contains ungrouped 13731 1727203821.28596: Examining possible inventory source: /tmp/network-zt6/inventory-rSl.yml 13731 1727203821.55200: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/cache 13731 1727203821.55260: Loading CacheModule 'memory' from /usr/local/lib/python3.12/site-packages/ansible/plugins/cache/memory.py 13731 1727203821.55488: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory 13731 1727203821.55546: Loading InventoryModule 'host_list' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/host_list.py 13731 1727203821.55621: Loaded config def from plugin (inventory/script) 13731 1727203821.55624: Loading InventoryModule 'script' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/script.py 13731 1727203821.55664: Loading InventoryModule 'auto' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/auto.py 13731 1727203821.55954: Loaded config def from plugin (inventory/yaml) 13731 1727203821.55956: Loading InventoryModule 'yaml' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/yaml.py 13731 1727203821.56043: Loading InventoryModule 'ini' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/ini.py 13731 1727203821.56858: Loading InventoryModule 'toml' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/toml.py 13731 1727203821.56862: Attempting to use plugin host_list (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/host_list.py) 13731 1727203821.56865: Attempting to use plugin script (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/script.py) 13731 1727203821.56871: Attempting to use plugin auto (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/auto.py) 13731 1727203821.57077: Loading data from /tmp/network-zt6/inventory-rSl.yml 13731 1727203821.57149: /tmp/network-zt6/inventory-rSl.yml was not parsable by auto 13731 1727203821.57213: Attempting to use plugin yaml (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/yaml.py) 13731 1727203821.57251: Loading data from /tmp/network-zt6/inventory-rSl.yml 13731 1727203821.57537: group all already in inventory 13731 1727203821.57544: set inventory_file for managed-node1 13731 1727203821.57547: set inventory_dir for managed-node1 13731 1727203821.57548: Added host managed-node1 to inventory 13731 1727203821.57550: Added host managed-node1 to group all 13731 1727203821.57551: set ansible_host for managed-node1 13731 1727203821.57552: set ansible_ssh_extra_args for managed-node1 13731 1727203821.57555: set inventory_file for managed-node2 13731 1727203821.57557: set inventory_dir for managed-node2 13731 1727203821.57558: Added host managed-node2 to inventory 13731 1727203821.57559: Added host managed-node2 to group all 13731 1727203821.57559: set ansible_host for managed-node2 13731 1727203821.57560: set ansible_ssh_extra_args for managed-node2 13731 1727203821.57563: set inventory_file for managed-node3 13731 1727203821.57565: set inventory_dir for managed-node3 13731 1727203821.57565: Added host managed-node3 to inventory 13731 1727203821.57566: Added host managed-node3 to group all 13731 1727203821.57567: set ansible_host for managed-node3 13731 1727203821.57568: set ansible_ssh_extra_args for managed-node3 13731 1727203821.57570: Reconcile groups and hosts in inventory. 13731 1727203821.57574: Group ungrouped now contains managed-node1 13731 1727203821.57577: Group ungrouped now contains managed-node2 13731 1727203821.57579: Group ungrouped now contains managed-node3 13731 1727203821.57653: '/usr/local/lib/python3.12/site-packages/ansible/plugins/vars/__init__' skipped due to reserved name 13731 1727203821.57974: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments 13731 1727203821.58023: Loading ModuleDocFragment 'vars_plugin_staging' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/vars_plugin_staging.py 13731 1727203821.58050: Loaded config def from plugin (vars/host_group_vars) 13731 1727203821.58053: Loading VarsModule 'host_group_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/vars/host_group_vars.py (found_in_cache=False, class_only=True) 13731 1727203821.58059: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/vars 13731 1727203821.58067: Loading VarsModule 'host_group_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/vars/host_group_vars.py (found_in_cache=True, class_only=False) 13731 1727203821.58313: Loading CacheModule 'memory' from /usr/local/lib/python3.12/site-packages/ansible/plugins/cache/memory.py (found_in_cache=True, class_only=False) 13731 1727203821.58840: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203821.59138: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py 13731 1727203821.59179: Loaded config def from plugin (connection/local) 13731 1727203821.59182: Loading Connection 'local' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/local.py (found_in_cache=False, class_only=True) 13731 1727203821.60185: Loaded config def from plugin (connection/paramiko_ssh) 13731 1727203821.60189: Loading Connection 'paramiko_ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/paramiko_ssh.py (found_in_cache=False, class_only=True) 13731 1727203821.61124: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 13731 1727203821.61161: Loaded config def from plugin (connection/psrp) 13731 1727203821.61164: Loading Connection 'psrp' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/psrp.py (found_in_cache=False, class_only=True) 13731 1727203821.61907: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 13731 1727203821.61951: Loaded config def from plugin (connection/ssh) 13731 1727203821.61954: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=False, class_only=True) 13731 1727203821.65263: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 13731 1727203821.65306: Loaded config def from plugin (connection/winrm) 13731 1727203821.65309: Loading Connection 'winrm' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/winrm.py (found_in_cache=False, class_only=True) 13731 1727203821.65341: '/usr/local/lib/python3.12/site-packages/ansible/plugins/shell/__init__' skipped due to reserved name 13731 1727203821.65607: Loading ModuleDocFragment 'shell_windows' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_windows.py 13731 1727203821.65672: Loaded config def from plugin (shell/cmd) 13731 1727203821.65674: Loading ShellModule 'cmd' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/cmd.py (found_in_cache=False, class_only=True) 13731 1727203821.65703: Loading ModuleDocFragment 'shell_windows' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_windows.py (found_in_cache=True, class_only=False) 13731 1727203821.65766: Loaded config def from plugin (shell/powershell) 13731 1727203821.65768: Loading ShellModule 'powershell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/powershell.py (found_in_cache=False, class_only=True) 13731 1727203821.66022: Loading ModuleDocFragment 'shell_common' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_common.py 13731 1727203821.66398: Loaded config def from plugin (shell/sh) 13731 1727203821.66400: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=False, class_only=True) 13731 1727203821.66434: '/usr/local/lib/python3.12/site-packages/ansible/plugins/become/__init__' skipped due to reserved name 13731 1727203821.66552: Loaded config def from plugin (become/runas) 13731 1727203821.66555: Loading BecomeModule 'runas' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/runas.py (found_in_cache=False, class_only=True) 13731 1727203821.66936: Loaded config def from plugin (become/su) 13731 1727203821.66939: Loading BecomeModule 'su' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/su.py (found_in_cache=False, class_only=True) 13731 1727203821.67296: Loaded config def from plugin (become/sudo) 13731 1727203821.67298: Loading BecomeModule 'sudo' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/sudo.py (found_in_cache=False, class_only=True) running playbook inside collection fedora.linux_system_roles 13731 1727203821.67331: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_options_nm.yml 13731 1727203821.68055: in VariableManager get_vars() 13731 1727203821.68079: done with get_vars() 13731 1727203821.68408: trying /usr/local/lib/python3.12/site-packages/ansible/modules 13731 1727203821.73882: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/action 13731 1727203821.73996: in VariableManager get_vars() 13731 1727203821.74002: done with get_vars() 13731 1727203821.74004: variable 'playbook_dir' from source: magic vars 13731 1727203821.74005: variable 'ansible_playbook_python' from source: magic vars 13731 1727203821.74006: variable 'ansible_config_file' from source: magic vars 13731 1727203821.74007: variable 'groups' from source: magic vars 13731 1727203821.74007: variable 'omit' from source: magic vars 13731 1727203821.74008: variable 'ansible_version' from source: magic vars 13731 1727203821.74009: variable 'ansible_check_mode' from source: magic vars 13731 1727203821.74010: variable 'ansible_diff_mode' from source: magic vars 13731 1727203821.74011: variable 'ansible_forks' from source: magic vars 13731 1727203821.74011: variable 'ansible_inventory_sources' from source: magic vars 13731 1727203821.74012: variable 'ansible_skip_tags' from source: magic vars 13731 1727203821.74013: variable 'ansible_limit' from source: magic vars 13731 1727203821.74013: variable 'ansible_run_tags' from source: magic vars 13731 1727203821.74014: variable 'ansible_verbosity' from source: magic vars 13731 1727203821.74050: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond_options.yml 13731 1727203821.74631: in VariableManager get_vars() 13731 1727203821.74649: done with get_vars() 13731 1727203821.74785: in VariableManager get_vars() 13731 1727203821.74798: done with get_vars() 13731 1727203821.74847: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/callback 13731 1727203821.74861: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__ redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug 13731 1727203821.75092: Loading ModuleDocFragment 'default_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/default_callback.py 13731 1727203821.75250: Loaded config def from plugin (callback/ansible_collections.ansible.posix.plugins.callback.debug) 13731 1727203821.75253: Loading CallbackModule 'ansible_collections.ansible.posix.plugins.callback.debug' from /tmp/collections-bGV/ansible_collections/ansible/posix/plugins/callback/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) 13731 1727203821.75286: '/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__init__' skipped due to reserved name 13731 1727203821.75311: Loading ModuleDocFragment 'default_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/default_callback.py (found_in_cache=True, class_only=False) 13731 1727203821.75480: Loading ModuleDocFragment 'result_format_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/result_format_callback.py 13731 1727203821.75546: Loaded config def from plugin (callback/default) 13731 1727203821.75548: Loading CallbackModule 'default' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/default.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) 13731 1727203821.76649: Loaded config def from plugin (callback/junit) 13731 1727203821.76652: Loading CallbackModule 'junit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/junit.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) 13731 1727203821.76696: Loading ModuleDocFragment 'result_format_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/result_format_callback.py (found_in_cache=True, class_only=False) 13731 1727203821.76757: Loaded config def from plugin (callback/minimal) 13731 1727203821.76759: Loading CallbackModule 'minimal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/minimal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) 13731 1727203821.76798: Loading CallbackModule 'oneline' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/oneline.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) 13731 1727203821.76857: Loaded config def from plugin (callback/tree) 13731 1727203821.76859: Loading CallbackModule 'tree' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/tree.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) redirecting (type: callback) ansible.builtin.profile_tasks to ansible.posix.profile_tasks 13731 1727203821.76971: Loaded config def from plugin (callback/ansible_collections.ansible.posix.plugins.callback.profile_tasks) 13731 1727203821.76974: Loading CallbackModule 'ansible_collections.ansible.posix.plugins.callback.profile_tasks' from /tmp/collections-bGV/ansible_collections/ansible/posix/plugins/callback/profile_tasks.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_bond_options_nm.yml ******************************************** 2 plays in /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_options_nm.yml 13731 1727203821.77007: in VariableManager get_vars() 13731 1727203821.77020: done with get_vars() 13731 1727203821.77026: in VariableManager get_vars() 13731 1727203821.77034: done with get_vars() 13731 1727203821.77038: variable 'omit' from source: magic vars 13731 1727203821.77079: in VariableManager get_vars() 13731 1727203821.77091: done with get_vars() 13731 1727203821.77110: variable 'omit' from source: magic vars PLAY [Run playbook 'playbooks/tests_bond_options.yml' with nm as provider] ***** 13731 1727203821.77909: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/strategy 13731 1727203821.81100: Loading StrategyModule 'linear' from /usr/local/lib/python3.12/site-packages/ansible/plugins/strategy/linear.py 13731 1727203821.81130: getting the remaining hosts for this loop 13731 1727203821.81132: done getting the remaining hosts for this loop 13731 1727203821.81135: getting the next task for host managed-node3 13731 1727203821.81138: done getting next task for host managed-node3 13731 1727203821.81140: ^ task is: TASK: Gathering Facts 13731 1727203821.81142: ^ state is: HOST STATE: block=0, task=0, rescue=0, always=0, handlers=0, run_state=0, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=True, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203821.81144: getting variables 13731 1727203821.81145: in VariableManager get_vars() 13731 1727203821.81155: Calling all_inventory to load vars for managed-node3 13731 1727203821.81158: Calling groups_inventory to load vars for managed-node3 13731 1727203821.81161: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203821.81173: Calling all_plugins_play to load vars for managed-node3 13731 1727203821.81185: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203821.81189: Calling groups_plugins_play to load vars for managed-node3 13731 1727203821.81221: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203821.81272: done with get_vars() 13731 1727203821.81281: done getting variables 13731 1727203821.81347: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (found_in_cache=False, class_only=True) TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_options_nm.yml:6 Tuesday 24 September 2024 14:50:21 -0400 (0:00:00.044) 0:00:00.044 ***** 13731 1727203821.81369: entering _queue_task() for managed-node3/gather_facts 13731 1727203821.81370: Creating lock for gather_facts 13731 1727203821.81694: worker is 1 (out of 1 available) 13731 1727203821.81706: exiting _queue_task() for managed-node3/gather_facts 13731 1727203821.81717: done queuing things up, now waiting for results queue to drain 13731 1727203821.81720: waiting for pending results... 13731 1727203821.82095: running TaskExecutor() for managed-node3/TASK: Gathering Facts 13731 1727203821.82101: in run() - task 028d2410-947f-82dc-c122-000000000015 13731 1727203821.82104: variable 'ansible_search_path' from source: unknown 13731 1727203821.82107: calling self._execute() 13731 1727203821.82149: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203821.82163: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203821.82178: variable 'omit' from source: magic vars 13731 1727203821.82279: variable 'omit' from source: magic vars 13731 1727203821.82310: variable 'omit' from source: magic vars 13731 1727203821.82353: variable 'omit' from source: magic vars 13731 1727203821.82401: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203821.82445: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203821.82468: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203821.82492: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203821.82508: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203821.82539: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203821.82552: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203821.82560: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203821.82664: Set connection var ansible_pipelining to False 13731 1727203821.82678: Set connection var ansible_shell_type to sh 13731 1727203821.82689: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203821.82698: Set connection var ansible_connection to ssh 13731 1727203821.82706: Set connection var ansible_shell_executable to /bin/sh 13731 1727203821.82714: Set connection var ansible_timeout to 10 13731 1727203821.82739: variable 'ansible_shell_executable' from source: unknown 13731 1727203821.82765: variable 'ansible_connection' from source: unknown 13731 1727203821.82768: variable 'ansible_module_compression' from source: unknown 13731 1727203821.82770: variable 'ansible_shell_type' from source: unknown 13731 1727203821.82772: variable 'ansible_shell_executable' from source: unknown 13731 1727203821.82776: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203821.82779: variable 'ansible_pipelining' from source: unknown 13731 1727203821.82781: variable 'ansible_timeout' from source: unknown 13731 1727203821.82876: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203821.82995: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (found_in_cache=True, class_only=False) 13731 1727203821.83011: variable 'omit' from source: magic vars 13731 1727203821.83020: starting attempt loop 13731 1727203821.83027: running the handler 13731 1727203821.83046: variable 'ansible_facts' from source: unknown 13731 1727203821.83068: _low_level_execute_command(): starting 13731 1727203821.83082: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203821.83800: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203821.83814: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203821.83826: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203821.83858: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203821.83957: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203821.83985: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203821.84057: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203821.85743: stdout chunk (state=3): >>>/root <<< 13731 1727203821.86108: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203821.86112: stdout chunk (state=3): >>><<< 13731 1727203821.86114: stderr chunk (state=3): >>><<< 13731 1727203821.86117: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203821.86119: _low_level_execute_command(): starting 13731 1727203821.86122: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604 `" && echo ansible-tmp-1727203821.860213-13800-179669116395604="` echo /root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604 `" ) && sleep 0' 13731 1727203821.87285: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203821.87380: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203821.87588: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203821.87592: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203821.87674: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203821.89620: stdout chunk (state=3): >>>ansible-tmp-1727203821.860213-13800-179669116395604=/root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604 <<< 13731 1727203821.89780: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203821.89784: stderr chunk (state=3): >>><<< 13731 1727203821.89786: stdout chunk (state=3): >>><<< 13731 1727203821.89828: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203821.860213-13800-179669116395604=/root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203821.90085: variable 'ansible_module_compression' from source: unknown 13731 1727203821.90088: ANSIBALLZ: Using generic lock for ansible.legacy.setup 13731 1727203821.90090: ANSIBALLZ: Acquiring lock 13731 1727203821.90092: ANSIBALLZ: Lock acquired: 140078454804688 13731 1727203821.90094: ANSIBALLZ: Creating module 13731 1727203822.44628: ANSIBALLZ: Writing module into payload 13731 1727203822.44779: ANSIBALLZ: Writing module 13731 1727203822.44807: ANSIBALLZ: Renaming module 13731 1727203822.44842: ANSIBALLZ: Done creating module 13731 1727203822.44887: variable 'ansible_facts' from source: unknown 13731 1727203822.44967: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203822.44984: _low_level_execute_command(): starting 13731 1727203822.45091: _low_level_execute_command(): executing: /bin/sh -c 'echo PLATFORM; uname; echo FOUND; command -v '"'"'python3.12'"'"'; command -v '"'"'python3.11'"'"'; command -v '"'"'python3.10'"'"'; command -v '"'"'python3.9'"'"'; command -v '"'"'python3.8'"'"'; command -v '"'"'python3.7'"'"'; command -v '"'"'/usr/bin/python3'"'"'; command -v '"'"'python3'"'"'; echo ENDFOUND && sleep 0' 13731 1727203822.46027: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203822.46044: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203822.46118: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 4 <<< 13731 1727203822.48369: stdout chunk (state=3): >>>PLATFORM <<< 13731 1727203822.48710: stdout chunk (state=3): >>>Linux FOUND /usr/bin/python3.12 <<< 13731 1727203822.48829: stdout chunk (state=3): >>>/usr/bin/python3 /usr/bin/python3 ENDFOUND <<< 13731 1727203822.48832: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203822.48835: stdout chunk (state=3): >>><<< 13731 1727203822.48837: stderr chunk (state=3): >>><<< 13731 1727203822.48839: _low_level_execute_command() done: rc=0, stdout=PLATFORM Linux FOUND /usr/bin/python3.12 /usr/bin/python3 /usr/bin/python3 ENDFOUND , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 4 debug2: Received exit status from master 0 13731 1727203822.48844 [managed-node3]: found interpreters: ['/usr/bin/python3.12', '/usr/bin/python3', '/usr/bin/python3'] 13731 1727203822.48970: _low_level_execute_command(): starting 13731 1727203822.48973: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 && sleep 0' 13731 1727203822.49110: Sending initial data 13731 1727203822.49120: Sent initial data (1181 bytes) 13731 1727203822.49894: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203822.49907: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203822.49918: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203822.49983: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203822.50068: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203822.50295: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203822.50393: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203822.53783: stdout chunk (state=3): >>>{"platform_dist_result": [], "osrelease_content": "NAME=\"CentOS Stream\"\nVERSION=\"10 (Coughlan)\"\nID=\"centos\"\nID_LIKE=\"rhel fedora\"\nVERSION_ID=\"10\"\nPLATFORM_ID=\"platform:el10\"\nPRETTY_NAME=\"CentOS Stream 10 (Coughlan)\"\nANSI_COLOR=\"0;31\"\nLOGO=\"fedora-logo-icon\"\nCPE_NAME=\"cpe:/o:centos:centos:10\"\nHOME_URL=\"https://centos.org/\"\nVENDOR_NAME=\"CentOS\"\nVENDOR_URL=\"https://centos.org/\"\nBUG_REPORT_URL=\"https://issues.redhat.com/\"\nREDHAT_SUPPORT_PRODUCT=\"Red Hat Enterprise Linux 10\"\nREDHAT_SUPPORT_PRODUCT_VERSION=\"CentOS Stream\"\n"} <<< 13731 1727203822.54390: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203822.54395: stderr chunk (state=3): >>><<< 13731 1727203822.54397: stdout chunk (state=3): >>><<< 13731 1727203822.54399: _low_level_execute_command() done: rc=0, stdout={"platform_dist_result": [], "osrelease_content": "NAME=\"CentOS Stream\"\nVERSION=\"10 (Coughlan)\"\nID=\"centos\"\nID_LIKE=\"rhel fedora\"\nVERSION_ID=\"10\"\nPLATFORM_ID=\"platform:el10\"\nPRETTY_NAME=\"CentOS Stream 10 (Coughlan)\"\nANSI_COLOR=\"0;31\"\nLOGO=\"fedora-logo-icon\"\nCPE_NAME=\"cpe:/o:centos:centos:10\"\nHOME_URL=\"https://centos.org/\"\nVENDOR_NAME=\"CentOS\"\nVENDOR_URL=\"https://centos.org/\"\nBUG_REPORT_URL=\"https://issues.redhat.com/\"\nREDHAT_SUPPORT_PRODUCT=\"Red Hat Enterprise Linux 10\"\nREDHAT_SUPPORT_PRODUCT_VERSION=\"CentOS Stream\"\n"} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203822.54402: variable 'ansible_facts' from source: unknown 13731 1727203822.54404: variable 'ansible_facts' from source: unknown 13731 1727203822.54406: variable 'ansible_module_compression' from source: unknown 13731 1727203822.54408: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.setup-ZIP_DEFLATED 13731 1727203822.54591: variable 'ansible_facts' from source: unknown 13731 1727203822.54974: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/AnsiballZ_setup.py 13731 1727203822.55385: Sending initial data 13731 1727203822.55394: Sent initial data (153 bytes) 13731 1727203822.56798: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203822.57015: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203822.57062: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203822.57143: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203822.58669: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203822.58705: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203822.58891: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpcbtpr4lr /root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/AnsiballZ_setup.py <<< 13731 1727203822.58895: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/AnsiballZ_setup.py" <<< 13731 1727203822.58928: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpcbtpr4lr" to remote "/root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/AnsiballZ_setup.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/AnsiballZ_setup.py" <<< 13731 1727203822.61695: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203822.61768: stderr chunk (state=3): >>><<< 13731 1727203822.61783: stdout chunk (state=3): >>><<< 13731 1727203822.61800: done transferring module to remote 13731 1727203822.61810: _low_level_execute_command(): starting 13731 1727203822.61815: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/ /root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/AnsiballZ_setup.py && sleep 0' 13731 1727203822.63182: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203822.63198: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203822.63202: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203822.63205: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203822.63208: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203822.63215: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203822.63330: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203822.63333: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203822.63527: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203822.65203: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203822.65305: stderr chunk (state=3): >>><<< 13731 1727203822.65309: stdout chunk (state=3): >>><<< 13731 1727203822.65326: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203822.65329: _low_level_execute_command(): starting 13731 1727203822.65371: _low_level_execute_command(): executing: /bin/sh -c 'PYTHONVERBOSE=1 /usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/AnsiballZ_setup.py && sleep 0' 13731 1727203822.66947: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203822.67042: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203822.67097: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203822.67200: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203822.67239: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203822.69613: stdout chunk (state=3): >>>import _frozen_importlib # frozen import _imp # builtin import '_thread' # import '_warnings' # import '_weakref' # import '_io' # import 'marshal' # <<< 13731 1727203822.69649: stdout chunk (state=3): >>>import 'posix' # <<< 13731 1727203822.69696: stdout chunk (state=3): >>>import '_frozen_importlib_external' # # installing zipimport hook <<< 13731 1727203822.69752: stdout chunk (state=3): >>>import 'time' # <<< 13731 1727203822.69755: stdout chunk (state=3): >>>import 'zipimport' # # installed zipimport hook <<< 13731 1727203822.69809: stdout chunk (state=3): >>># /usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/encodings/__init__.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.69855: stdout chunk (state=3): >>>import '_codecs' # import 'codecs' # <<< 13731 1727203822.69984: stdout chunk (state=3): >>># /usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc matches /usr/lib64/python3.12/encodings/aliases.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79774184d0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79773e7b30> # /usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc' import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797741aa50> <<< 13731 1727203822.70008: stdout chunk (state=3): >>>import '_signal' # <<< 13731 1727203822.70065: stdout chunk (state=3): >>>import '_abc' # <<< 13731 1727203822.70078: stdout chunk (state=3): >>>import 'abc' # import 'io' # <<< 13731 1727203822.70310: stdout chunk (state=3): >>>import '_stat' # import 'stat' # import '_collections_abc' # import 'genericpath' # import 'posixpath' # <<< 13731 1727203822.70318: stdout chunk (state=3): >>>import 'os' # <<< 13731 1727203822.70324: stdout chunk (state=3): >>>import '_sitebuiltins' # <<< 13731 1727203822.70351: stdout chunk (state=3): >>>Processing user site-packages Processing global site-packages Adding directory: '/usr/local/lib/python3.12/site-packages' <<< 13731 1727203822.70459: stdout chunk (state=3): >>>Adding directory: '/usr/lib64/python3.12/site-packages' Adding directory: '/usr/lib/python3.12/site-packages' Processing .pth file: '/usr/lib/python3.12/site-packages/distutils-precedence.pth' # /usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8_sig.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc' import 'encodings.utf_8_sig' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977209130> <<< 13731 1727203822.70512: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/_distutils_hack/__init__.py <<< 13731 1727203822.70565: stdout chunk (state=3): >>># code object from '/usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc' import '_distutils_hack' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797720a060> <<< 13731 1727203822.70587: stdout chunk (state=3): >>>import 'site' # <<< 13731 1727203822.70715: stdout chunk (state=3): >>>Python 3.12.5 (main, Aug 23 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-1)] on linux Type "help", "copyright", "credits" or "license" for more information. <<< 13731 1727203822.71284: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc matches /usr/lib64/python3.12/base64.py <<< 13731 1727203822.71306: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc' <<< 13731 1727203822.71417: stdout chunk (state=3): >>># /usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/re/__init__.py # code object from '/usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc matches /usr/lib64/python3.12/enum.py # code object from '/usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc' <<< 13731 1727203822.71633: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/types.cpython-312.pyc matches /usr/lib64/python3.12/types.py # code object from '/usr/lib64/python3.12/__pycache__/types.cpython-312.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977247e90> # /usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc matches /usr/lib64/python3.12/operator.py # code object from '/usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc' <<< 13731 1727203822.71717: stdout chunk (state=3): >>>import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977247f50> # /usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc matches /usr/lib64/python3.12/functools.py # code object from '/usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc' # /usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/collections/__init__.py # code object from '/usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.71854: stdout chunk (state=3): >>>import 'itertools' # # /usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc matches /usr/lib64/python3.12/keyword.py # code object from '/usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797727f890> # /usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc matches /usr/lib64/python3.12/reprlib.py # code object from '/usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797727ff20> import '_collections' # <<< 13731 1727203822.71858: stdout chunk (state=3): >>>import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797725fb60> <<< 13731 1727203822.71918: stdout chunk (state=3): >>>import '_functools' # <<< 13731 1727203822.71921: stdout chunk (state=3): >>>import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797725d280> <<< 13731 1727203822.72129: stdout chunk (state=3): >>>import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977245040> # /usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc matches /usr/lib64/python3.12/re/_compiler.py # code object from '/usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc' import '_sre' # <<< 13731 1727203822.72280: stdout chunk (state=3): >>># /usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc matches /usr/lib64/python3.12/re/_parser.py # code object from '/usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc' <<< 13731 1727203822.72310: stdout chunk (state=3): >>># /usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc matches /usr/lib64/python3.12/re/_constants.py # code object from '/usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc' import 're._constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772a3770> <<< 13731 1727203822.72417: stdout chunk (state=3): >>>import 're._parser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772a2390> # /usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc matches /usr/lib64/python3.12/re/_casefix.py # code object from '/usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc' import 're._casefix' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797725e120> import 're._compiler' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977246900> # /usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc matches /usr/lib64/python3.12/copyreg.py # code object from '/usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc' <<< 13731 1727203822.72433: stdout chunk (state=3): >>>import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d4830> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772442c0> # /usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc matches /usr/lib64/python3.12/struct.py # code object from '/usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772d4ce0> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d4b90> <<< 13731 1727203822.72453: stdout chunk (state=3): >>># extension module 'binascii' loaded from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772d4f80> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977242de0> <<< 13731 1727203822.72673: stdout chunk (state=3): >>># /usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/importlib/__init__.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.72679: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc matches /usr/lib64/python3.12/warnings.py <<< 13731 1727203822.72698: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d5640> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d5310> import 'importlib.machinery' # # /usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc matches /usr/lib64/python3.12/importlib/_abc.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc' import 'importlib._abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d6510> import 'importlib.util' # import 'runpy' # # /usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc matches /usr/lib64/python3.12/shutil.py # code object from '/usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc' <<< 13731 1727203822.72736: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc matches /usr/lib64/python3.12/fnmatch.py # code object from '/usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772ec710> import 'errno' # <<< 13731 1727203822.72770: stdout chunk (state=3): >>># extension module 'zlib' loaded from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.72807: stdout chunk (state=3): >>>import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772eddc0> # /usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc matches /usr/lib64/python3.12/bz2.py # code object from '/usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc' <<< 13731 1727203822.72913: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc matches /usr/lib64/python3.12/_compression.py # code object from '/usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc' <<< 13731 1727203822.72967: stdout chunk (state=3): >>>import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772eec60> # extension module '_bz2' loaded from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772ef290> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772ee1b0> # /usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc matches /usr/lib64/python3.12/lzma.py # code object from '/usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc' <<< 13731 1727203822.73006: stdout chunk (state=3): >>># extension module '_lzma' loaded from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772efd10> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772ef440> <<< 13731 1727203822.73113: stdout chunk (state=3): >>>import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d6480> <<< 13731 1727203822.73118: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc matches /usr/lib64/python3.12/tempfile.py <<< 13731 1727203822.73148: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/random.cpython-312.pyc matches /usr/lib64/python3.12/random.py # code object from '/usr/lib64/python3.12/__pycache__/random.cpython-312.pyc' # extension module 'math' loaded from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976ff7c80> <<< 13731 1727203822.73172: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc matches /usr/lib64/python3.12/bisect.py # code object from '/usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc' <<< 13731 1727203822.73200: stdout chunk (state=3): >>># extension module '_bisect' loaded from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79770206b0> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977020440> <<< 13731 1727203822.73547: stdout chunk (state=3): >>># extension module '_random' loaded from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79770206e0> # /usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc matches /usr/lib64/python3.12/hashlib.py # code object from '/usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc' # extension module '_hashlib' loaded from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.73597: stdout chunk (state=3): >>># extension module '_hashlib' executed from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7977021010> # extension module '_blake2' loaded from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7977021970> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79770208c0> <<< 13731 1727203822.73625: stdout chunk (state=3): >>>import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976ff5e20> # /usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc matches /usr/lib64/python3.12/weakref.py <<< 13731 1727203822.73651: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc' <<< 13731 1727203822.73706: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc matches /usr/lib64/python3.12/_weakrefset.py <<< 13731 1727203822.73752: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977022d20> import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977020e90> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d6c30> # /usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/__init__.py <<< 13731 1727203822.74048: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc matches /usr/lib64/python3.12/threading.py # code object from '/usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc' import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797704f080> # /usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/__init__.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc matches /usr/lib64/python3.12/contextlib.py <<< 13731 1727203822.74051: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc' <<< 13731 1727203822.74115: stdout chunk (state=3): >>>import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977073410> # /usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc matches /usr/lib64/python3.12/pathlib.py <<< 13731 1727203822.74171: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc' <<< 13731 1727203822.74292: stdout chunk (state=3): >>>import 'ntpath' # # /usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/urllib/__init__.py # code object from '/usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc' import 'urllib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79770d01d0> <<< 13731 1727203822.74303: stdout chunk (state=3): >>># /usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc matches /usr/lib64/python3.12/urllib/parse.py <<< 13731 1727203822.74366: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc matches /usr/lib64/python3.12/ipaddress.py <<< 13731 1727203822.74424: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc' <<< 13731 1727203822.74542: stdout chunk (state=3): >>>import 'ipaddress' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79770d2930> <<< 13731 1727203822.74646: stdout chunk (state=3): >>>import 'urllib.parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79770d02f0> <<< 13731 1727203822.74739: stdout chunk (state=3): >>>import 'pathlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797709d1f0> # /usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/glob.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc' import 'zipfile._path.glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797709d9a0> <<< 13731 1727203822.74754: stdout chunk (state=3): >>>import 'zipfile._path' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977072210> import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977023c50> <<< 13731 1727203822.75251: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/encodings/cp437.pyc' import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f7977072330> <<< 13731 1727203822.75729: stdout chunk (state=3): >>># zipimport: found 103 names in '/tmp/ansible_ansible.legacy.setup_payload_r1p_k8g4/ansible_ansible.legacy.setup_payload.zip' # zipimport: zlib available<<< 13731 1727203822.76025: stdout chunk (state=3): >>> # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc matches /usr/lib64/python3.12/pkgutil.py # code object from '/usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc' <<< 13731 1727203822.76065: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc matches /usr/lib64/python3.12/typing.py <<< 13731 1727203822.76180: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc' <<< 13731 1727203822.76236: stdout chunk (state=3): >>># /usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc matches /usr/lib64/python3.12/collections/abc.py <<< 13731 1727203822.76265: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797698aff0> import '_typing' # <<< 13731 1727203822.76571: stdout chunk (state=3): >>>import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976969ee0> import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976969070> # zipimport: zlib available<<< 13731 1727203822.76639: stdout chunk (state=3): >>> import 'ansible' # <<< 13731 1727203822.76672: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.76701: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.76727: stdout chunk (state=3): >>>import 'ansible.module_utils' # <<< 13731 1727203822.76767: stdout chunk (state=3): >>> # zipimport: zlib available <<< 13731 1727203822.78999: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.80317: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc matches /usr/lib64/python3.12/__future__.py # code object from '/usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976988e90> <<< 13731 1727203822.80343: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/json/__init__.py # code object from '/usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.80350: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc matches /usr/lib64/python3.12/json/decoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc' <<< 13731 1727203822.80398: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc matches /usr/lib64/python3.12/json/scanner.py # code object from '/usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc' <<< 13731 1727203822.80402: stdout chunk (state=3): >>># extension module '_json' loaded from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79769c2960> <<< 13731 1727203822.80431: stdout chunk (state=3): >>>import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79769c26f0> <<< 13731 1727203822.80471: stdout chunk (state=3): >>>import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79769c2000> <<< 13731 1727203822.80496: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc matches /usr/lib64/python3.12/json/encoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc' <<< 13731 1727203822.80534: stdout chunk (state=3): >>>import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79769c2a50> <<< 13731 1727203822.80618: stdout chunk (state=3): >>>import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797698ba10> import 'atexit' # # extension module 'grp' loaded from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79769c36e0> <<< 13731 1727203822.80621: stdout chunk (state=3): >>># extension module 'fcntl' loaded from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79769c3920> # /usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc matches /usr/lib64/python3.12/locale.py <<< 13731 1727203822.80712: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc' <<< 13731 1727203822.80838: stdout chunk (state=3): >>>import '_locale' # import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79769c3e30> import 'pwd' # <<< 13731 1727203822.80841: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc matches /usr/lib64/python3.12/platform.py <<< 13731 1727203822.80858: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc' import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976825b20> # extension module 'select' loaded from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79768277d0> # /usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc matches /usr/lib64/python3.12/selectors.py <<< 13731 1727203822.80886: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc' <<< 13731 1727203822.80920: stdout chunk (state=3): >>>import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682c1a0> <<< 13731 1727203822.80953: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc matches /usr/lib64/python3.12/shlex.py <<< 13731 1727203822.80986: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc' <<< 13731 1727203822.81010: stdout chunk (state=3): >>>import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682d340> # /usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc matches /usr/lib64/python3.12/subprocess.py <<< 13731 1727203822.81018: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc' <<< 13731 1727203822.81085: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc matches /usr/lib64/python3.12/signal.py # code object from '/usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc' <<< 13731 1727203822.81118: stdout chunk (state=3): >>>import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682fda0> <<< 13731 1727203822.81186: stdout chunk (state=3): >>># extension module '_posixsubprocess' loaded from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797682fe30> import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682e090> <<< 13731 1727203822.81231: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc matches /usr/lib64/python3.12/traceback.py # code object from '/usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc' <<< 13731 1727203822.81257: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc matches /usr/lib64/python3.12/linecache.py # code object from '/usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc matches /usr/lib64/python3.12/tokenize.py <<< 13731 1727203822.81365: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/token.cpython-312.pyc matches /usr/lib64/python3.12/token.py # code object from '/usr/lib64/python3.12/__pycache__/token.cpython-312.pyc' <<< 13731 1727203822.81383: stdout chunk (state=3): >>>import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976833c80> <<< 13731 1727203822.81393: stdout chunk (state=3): >>>import '_tokenize' # <<< 13731 1727203822.81660: stdout chunk (state=3): >>>import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976832780> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768324e0> # /usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc matches /usr/lib64/python3.12/textwrap.py # code object from '/usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc' import 'textwrap' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976832a50> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682e570> # extension module 'syslog' loaded from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976877f80> <<< 13731 1727203822.81715: stdout chunk (state=3): >>># /usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768784a0> <<< 13731 1727203822.81727: stdout chunk (state=3): >>># /usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/journal.py <<< 13731 1727203822.81894: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc matches /usr/lib64/python3.12/datetime.py # code object from '/usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc' # extension module '_datetime' loaded from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976879af0> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768798b0> # /usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc matches /usr/lib64/python3.12/uuid.py # code object from '/usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc' <<< 13731 1727203822.81928: stdout chunk (state=3): >>># extension module '_uuid' loaded from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.81942: stdout chunk (state=3): >>># extension module '_uuid' executed from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797687bfe0> <<< 13731 1727203822.81951: stdout chunk (state=3): >>>import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797687a1b0> <<< 13731 1727203822.81989: stdout chunk (state=3): >>># /usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/logging/__init__.py <<< 13731 1727203822.82041: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.82071: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/string.cpython-312.pyc matches /usr/lib64/python3.12/string.py <<< 13731 1727203822.82090: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/string.cpython-312.pyc' <<< 13731 1727203822.82112: stdout chunk (state=3): >>>import '_string' # <<< 13731 1727203822.82176: stdout chunk (state=3): >>>import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797687f830> <<< 13731 1727203822.82371: stdout chunk (state=3): >>>import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797687c200> <<< 13731 1727203822.82448: stdout chunk (state=3): >>># extension module 'systemd._journal' loaded from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.82462: stdout chunk (state=3): >>># extension module 'systemd._journal' executed from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79768805f0> <<< 13731 1727203822.82499: stdout chunk (state=3): >>># extension module 'systemd._reader' loaded from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.82562: stdout chunk (state=3): >>># extension module 'systemd._reader' executed from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976880650> <<< 13731 1727203822.82584: stdout chunk (state=3): >>># extension module 'systemd.id128' loaded from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.82590: stdout chunk (state=3): >>># extension module 'systemd.id128' executed from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.82693: stdout chunk (state=3): >>>import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976880bf0> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976878200> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc matches /usr/lib64/python3.12/socket.py # code object from '/usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc' <<< 13731 1727203822.82717: stdout chunk (state=3): >>># extension module '_socket' loaded from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.82753: stdout chunk (state=3): >>># extension module '_socket' executed from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.82765: stdout chunk (state=3): >>>import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797670c350> <<< 13731 1727203822.82995: stdout chunk (state=3): >>># extension module 'array' loaded from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.83014: stdout chunk (state=3): >>># extension module 'array' executed from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203822.83027: stdout chunk (state=3): >>> import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797670d670> <<< 13731 1727203822.83043: stdout chunk (state=3): >>>import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976882ae0> <<< 13731 1727203822.83078: stdout chunk (state=3): >>># extension module 'systemd._daemon' loaded from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.83091: stdout chunk (state=3): >>># extension module 'systemd._daemon' executed from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976883e60> <<< 13731 1727203822.83196: stdout chunk (state=3): >>>import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768826c0> # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat' # # zipimport: zlib available <<< 13731 1727203822.83311: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.83482: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.83485: stdout chunk (state=3): >>>import 'ansible.module_utils.common' # <<< 13731 1727203822.83488: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.83513: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.83541: stdout chunk (state=3): >>>import 'ansible.module_utils.common.text' # <<< 13731 1727203822.83554: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.83742: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.83932: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.84643: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.85193: stdout chunk (state=3): >>>import 'ansible.module_utils.six' # import 'ansible.module_utils.six.moves' # <<< 13731 1727203822.85197: stdout chunk (state=3): >>>import 'ansible.module_utils.six.moves.collections_abc' # import 'ansible.module_utils.common.text.converters' # <<< 13731 1727203822.85231: stdout chunk (state=3): >>># /usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/__init__.py <<< 13731 1727203822.85240: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.85447: stdout chunk (state=3): >>># extension module '_ctypes' loaded from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976711760> # /usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/_endian.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc' import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976712510> <<< 13731 1727203822.85450: stdout chunk (state=3): >>>import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797670d850> <<< 13731 1727203822.85521: stdout chunk (state=3): >>>import 'ansible.module_utils.compat.selinux' # # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.85552: stdout chunk (state=3): >>>import 'ansible.module_utils._text' # # zipimport: zlib available <<< 13731 1727203822.85767: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.86008: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc matches /usr/lib64/python3.12/copy.py # code object from '/usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc' import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767124b0> <<< 13731 1727203822.86085: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.86791: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.87487: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.87594: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.87746: stdout chunk (state=3): >>>import 'ansible.module_utils.common.collections' # # zipimport: zlib available <<< 13731 1727203822.87753: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.87800: stdout chunk (state=3): >>>import 'ansible.module_utils.common.warnings' # <<< 13731 1727203822.87877: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.87928: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.88026: stdout chunk (state=3): >>>import 'ansible.module_utils.errors' # <<< 13731 1727203822.88039: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.parsing' # <<< 13731 1727203822.88056: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.88115: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.88139: stdout chunk (state=3): >>>import 'ansible.module_utils.parsing.convert_bool' # # zipimport: zlib available <<< 13731 1727203822.88380: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.88780: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc matches /usr/lib64/python3.12/ast.py <<< 13731 1727203822.88784: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc' import '_ast' # import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976713680> # zipimport: zlib available <<< 13731 1727203822.88829: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.88906: stdout chunk (state=3): >>>import 'ansible.module_utils.common.text.formatters' # import 'ansible.module_utils.common.validation' # <<< 13731 1727203822.89350: stdout chunk (state=3): >>>import 'ansible.module_utils.common.parameters' # import 'ansible.module_utils.common.arg_spec' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.locale' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/selinux/__init__.py <<< 13731 1727203822.89395: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.89683: stdout chunk (state=3): >>># extension module 'selinux._selinux' loaded from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797671e030> <<< 13731 1727203822.89716: stdout chunk (state=3): >>>import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976719040> import 'ansible.module_utils.common.file' # import 'ansible.module_utils.common.process' # # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.89768: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.89809: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.89853: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/__init__.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.89902: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/distro.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc' <<< 13731 1727203822.89942: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc matches /usr/lib64/python3.12/argparse.py <<< 13731 1727203822.90020: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc matches /usr/lib64/python3.12/gettext.py <<< 13731 1727203822.90115: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc' <<< 13731 1727203822.90181: stdout chunk (state=3): >>>import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976806ae0> <<< 13731 1727203822.90194: stdout chunk (state=3): >>>import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768fe7b0> <<< 13731 1727203822.90301: stdout chunk (state=3): >>>import 'distro.distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797671e270> import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976714890> # destroy ansible.module_utils.distro import 'ansible.module_utils.distro' # <<< 13731 1727203822.90337: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.90383: stdout chunk (state=3): >>>import 'ansible.module_utils.common._utils' # import 'ansible.module_utils.common.sys_info' # <<< 13731 1727203822.90466: stdout chunk (state=3): >>>import 'ansible.module_utils.basic' # # zipimport: zlib available <<< 13731 1727203822.90510: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.modules' # # zipimport: zlib available <<< 13731 1727203822.90578: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.90824: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.90850: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.90921: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.namespace' # <<< 13731 1727203822.90954: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.91039: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.91168: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.91218: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.91223: stdout chunk (state=3): >>>import 'ansible.module_utils.compat.typing' # # zipimport: zlib available <<< 13731 1727203822.91521: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.91788: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.91883: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/__init__.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203822.91905: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/context.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/context.py <<< 13731 1727203822.91911: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/context.cpython-312.pyc' <<< 13731 1727203822.91997: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/process.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/process.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/process.cpython-312.pyc' <<< 13731 1727203822.92001: stdout chunk (state=3): >>>import 'multiprocessing.process' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b23f0> <<< 13731 1727203822.92063: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/reduction.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/reduction.py <<< 13731 1727203822.92066: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/reduction.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/pickle.cpython-312.pyc matches /usr/lib64/python3.12/pickle.py <<< 13731 1727203822.92091: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/pickle.cpython-312.pyc' <<< 13731 1727203822.92144: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/_compat_pickle.cpython-312.pyc matches /usr/lib64/python3.12/_compat_pickle.py # code object from '/usr/lib64/python3.12/__pycache__/_compat_pickle.cpython-312.pyc' import '_compat_pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976420110> <<< 13731 1727203822.92239: stdout chunk (state=3): >>># extension module '_pickle' loaded from '/usr/lib64/python3.12/lib-dynload/_pickle.cpython-312-x86_64-linux-gnu.so' # extension module '_pickle' executed from '/usr/lib64/python3.12/lib-dynload/_pickle.cpython-312-x86_64-linux-gnu.so' import '_pickle' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79764206e0> <<< 13731 1727203822.92287: stdout chunk (state=3): >>>import 'pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797679f050> import 'multiprocessing.reduction' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b2f90> <<< 13731 1727203822.92418: stdout chunk (state=3): >>>import 'multiprocessing.context' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b0aa0> import 'multiprocessing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b13a0> # /usr/lib64/python3.12/multiprocessing/__pycache__/pool.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/pool.py <<< 13731 1727203822.92461: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/pool.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/queue.cpython-312.pyc matches /usr/lib64/python3.12/queue.py # code object from '/usr/lib64/python3.12/__pycache__/queue.cpython-312.pyc' <<< 13731 1727203822.92514: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/heapq.cpython-312.pyc matches /usr/lib64/python3.12/heapq.py # code object from '/usr/lib64/python3.12/__pycache__/heapq.cpython-312.pyc' <<< 13731 1727203822.92557: stdout chunk (state=3): >>># extension module '_heapq' loaded from '/usr/lib64/python3.12/lib-dynload/_heapq.cpython-312-x86_64-linux-gnu.so' # extension module '_heapq' executed from '/usr/lib64/python3.12/lib-dynload/_heapq.cpython-312-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79764233e0> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976422c90> # extension module '_queue' loaded from '/usr/lib64/python3.12/lib-dynload/_queue.cpython-312-x86_64-linux-gnu.so' # extension module '_queue' executed from '/usr/lib64/python3.12/lib-dynload/_queue.cpython-312-x86_64-linux-gnu.so' import '_queue' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976422e70> <<< 13731 1727203822.92588: stdout chunk (state=3): >>>import 'queue' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79764220f0> # /usr/lib64/python3.12/multiprocessing/__pycache__/util.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/util.py <<< 13731 1727203822.92904: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/util.cpython-312.pyc' import 'multiprocessing.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976423530> # /usr/lib64/python3.12/multiprocessing/__pycache__/connection.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/connection.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/connection.cpython-312.pyc' # extension module '_multiprocessing' loaded from '/usr/lib64/python3.12/lib-dynload/_multiprocessing.cpython-312-x86_64-linux-gnu.so' # extension module '_multiprocessing' executed from '/usr/lib64/python3.12/lib-dynload/_multiprocessing.cpython-312-x86_64-linux-gnu.so' import '_multiprocessing' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976486060> import 'multiprocessing.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976423f80> import 'multiprocessing.pool' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b1550> import 'ansible.module_utils.facts.timeout' # import 'ansible.module_utils.facts.collector' # <<< 13731 1727203822.92946: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.other' # # zipimport: zlib available <<< 13731 1727203822.93095: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.other.facter' # <<< 13731 1727203822.93114: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.93180: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.93281: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.other.ohai' # <<< 13731 1727203822.93285: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system' # <<< 13731 1727203822.93287: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.93357: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.system.apparmor' # # zipimport: zlib available <<< 13731 1727203822.93427: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.93495: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.caps' # # zipimport: zlib available <<< 13731 1727203822.93550: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.93606: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.chroot' # # zipimport: zlib available <<< 13731 1727203822.93694: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.93790: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.93853: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.93936: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.utils' # import 'ansible.module_utils.facts.system.cmdline' # # zipimport: zlib available <<< 13731 1727203822.94798: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95371: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.distribution' # <<< 13731 1727203822.95383: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95448: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95525: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95564: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95615: stdout chunk (state=3): >>>import 'ansible.module_utils.compat.datetime' # import 'ansible.module_utils.facts.system.date_time' # <<< 13731 1727203822.95621: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95658: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95703: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.env' # <<< 13731 1727203822.95706: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95779: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95849: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.dns' # <<< 13731 1727203822.95890: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95908: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.95954: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.fips' # # zipimport: zlib available <<< 13731 1727203822.96028: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.system.loadavg' # <<< 13731 1727203822.96040: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.96144: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.96266: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/glob.py # code object from '/usr/lib64/python3.12/__pycache__/glob.cpython-312.pyc' <<< 13731 1727203822.96309: stdout chunk (state=3): >>>import 'glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976487830> # /usr/lib64/python3.12/__pycache__/configparser.cpython-312.pyc matches /usr/lib64/python3.12/configparser.py <<< 13731 1727203822.96345: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/configparser.cpython-312.pyc' <<< 13731 1727203822.96523: stdout chunk (state=3): >>>import 'configparser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976486bd0> import 'ansible.module_utils.facts.system.local' # <<< 13731 1727203822.96625: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.96713: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.lsb' # <<< 13731 1727203822.96728: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.96853: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.96986: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.pkg_mgr' # <<< 13731 1727203822.96993: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.97086: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.97182: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.platform' # <<< 13731 1727203822.97197: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.97248: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.97366: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/ssl.cpython-312.pyc matches /usr/lib64/python3.12/ssl.py # code object from '/usr/lib64/python3.12/__pycache__/ssl.cpython-312.pyc' <<< 13731 1727203822.97450: stdout chunk (state=3): >>># extension module '_ssl' loaded from '/usr/lib64/python3.12/lib-dynload/_ssl.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.97536: stdout chunk (state=3): >>># extension module '_ssl' executed from '/usr/lib64/python3.12/lib-dynload/_ssl.cpython-312-x86_64-linux-gnu.so' import '_ssl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79764be270> <<< 13731 1727203822.97816: stdout chunk (state=3): >>>import 'ssl' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79764af140> <<< 13731 1727203822.97822: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.python' # <<< 13731 1727203822.97888: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.97909: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.97987: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.selinux' # <<< 13731 1727203822.97993: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.98107: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.98370: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.98584: stdout chunk (state=3): >>>import 'ansible.module_utils.compat.version' # <<< 13731 1727203822.98591: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.service_mgr' # # zipimport: zlib available <<< 13731 1727203822.98647: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.98701: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.ssh_pub_keys' # <<< 13731 1727203822.98709: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.98787: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.98818: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/getpass.cpython-312.pyc matches /usr/lib64/python3.12/getpass.py <<< 13731 1727203822.98824: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/getpass.cpython-312.pyc' <<< 13731 1727203822.98857: stdout chunk (state=3): >>># extension module 'termios' loaded from '/usr/lib64/python3.12/lib-dynload/termios.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203822.98877: stdout chunk (state=3): >>># extension module 'termios' executed from '/usr/lib64/python3.12/lib-dynload/termios.cpython-312-x86_64-linux-gnu.so' import 'termios' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79764d1dc0> import 'getpass' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79764be060> <<< 13731 1727203822.98903: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.user' # # zipimport: zlib available <<< 13731 1727203822.98914: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.hardware' # <<< 13731 1727203822.98982: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.99037: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.base' # <<< 13731 1727203822.99043: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.99274: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.99498: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.aix' # <<< 13731 1727203822.99656: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203822.99807: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.99858: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.99911: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.sysctl' # import 'ansible.module_utils.facts.hardware.darwin' # <<< 13731 1727203822.99932: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.99952: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203822.99981: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.00190: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.00401: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.freebsd' # import 'ansible.module_utils.facts.hardware.dragonfly' # <<< 13731 1727203823.00415: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.00599: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.00785: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.hpux' # <<< 13731 1727203823.00791: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.00834: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.00878: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.01728: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.02518: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.linux' # <<< 13731 1727203823.02524: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.hurd' # <<< 13731 1727203823.02588: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.02688: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.02844: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.netbsd' # <<< 13731 1727203823.02851: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.02994: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.03145: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.openbsd' # <<< 13731 1727203823.03288: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.03395: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.03630: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.sunos' # <<< 13731 1727203823.03641: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.03662: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.network' # <<< 13731 1727203823.03679: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.03731: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.03793: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.base' # <<< 13731 1727203823.03799: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.03942: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.04090: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.04402: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.04736: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.generic_bsd' # <<< 13731 1727203823.04748: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.aix' # <<< 13731 1727203823.04758: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.04796: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.04845: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.darwin' # <<< 13731 1727203823.04884: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203823.04917: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.dragonfly' # <<< 13731 1727203823.04922: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.05122: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.network.fc_wwn' # <<< 13731 1727203823.05127: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.05160: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.05195: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.freebsd' # <<< 13731 1727203823.05202: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.05274: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.05358: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.hpux' # <<< 13731 1727203823.05366: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.05446: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.05520: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.hurd' # <<< 13731 1727203823.05692: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.05953: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.06369: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.linux' # <<< 13731 1727203823.06378: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.06457: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.06693: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.iscsi' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.nvme' # # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203823.06717: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.netbsd' # <<< 13731 1727203823.06723: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.06765: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.06799: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.openbsd' # <<< 13731 1727203823.06821: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.06932: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.07048: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.sunos' # <<< 13731 1727203823.07060: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.07076: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.virtual' # <<< 13731 1727203823.07098: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.07154: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.07215: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.base' # <<< 13731 1727203823.07220: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.07271: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203823.07334: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.07493: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203823.07609: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.sysctl' # <<< 13731 1727203823.07613: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.freebsd' # import 'ansible.module_utils.facts.virtual.dragonfly' # <<< 13731 1727203823.07630: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.07703: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.07772: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.hpux' # <<< 13731 1727203823.07780: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.08079: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.08387: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.linux' # <<< 13731 1727203823.08393: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.08492: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.08521: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.netbsd' # <<< 13731 1727203823.08530: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.08612: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.08694: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.openbsd' # # zipimport: zlib available <<< 13731 1727203823.08793: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.08900: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.sunos' # import 'ansible.module_utils.facts.default_collectors' # <<< 13731 1727203823.08999: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.09043: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.09170: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.ansible_collector' # import 'ansible.module_utils.facts.compat' # import 'ansible.module_utils.facts' # <<< 13731 1727203823.09274: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203823.09768: stdout chunk (state=3): >>># /usr/lib64/python3.12/encodings/__pycache__/idna.cpython-312.pyc matches /usr/lib64/python3.12/encodings/idna.py <<< 13731 1727203823.09774: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/encodings/__pycache__/idna.cpython-312.pyc' <<< 13731 1727203823.09798: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/stringprep.cpython-312.pyc matches /usr/lib64/python3.12/stringprep.py <<< 13731 1727203823.09817: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/stringprep.cpython-312.pyc' <<< 13731 1727203823.09850: stdout chunk (state=3): >>># extension module 'unicodedata' loaded from '/usr/lib64/python3.12/lib-dynload/unicodedata.cpython-312-x86_64-linux-gnu.so' # extension module 'unicodedata' executed from '/usr/lib64/python3.12/lib-dynload/unicodedata.cpython-312-x86_64-linux-gnu.so' import 'unicodedata' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976273860> <<< 13731 1727203823.09939: stdout chunk (state=3): >>>import 'stringprep' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976270740> import 'encodings.idna' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976273140> <<< 13731 1727203823.22383: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/queues.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/queues.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/queues.cpython-312.pyc' import 'multiprocessing.queues' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79762b87a0> <<< 13731 1727203823.22387: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/synchronize.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/synchronize.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/synchronize.cpython-312.pyc' import 'multiprocessing.synchronize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79762b96a0> <<< 13731 1727203823.22390: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/dummy/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/dummy/__init__.py # code object from '/usr/lib64/python3.12/multiprocessing/dummy/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203823.22392: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/dummy/__pycache__/connection.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/dummy/connection.py # code object from '/usr/lib64/python3.12/multiprocessing/dummy/__pycache__/connection.cpython-312.pyc' import 'multiprocessing.dummy.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976306420> <<< 13731 1727203823.22394: stdout chunk (state=3): >>>import 'multiprocessing.dummy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79762bb4d0> <<< 13731 1727203823.22565: stdout chunk (state=3): >>>PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame <<< 13731 1727203823.47531: stdout chunk (state=3): >>> {"ansible_facts": {"ansible_env": {"PYTHONVERBOSE": "1", "SHELL": "/bin/bash", "GPG_TTY": "/dev/pts/0", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.12", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.45.138 53526 10.31.47.22 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.45.138 53526 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0"}, "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "Super User", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_system": "Linux", "ansible_kernel": "6.11.0-25.el10.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Mon Sep 16 20:35:26 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.12.5", "ansible_fqdn": "managed-node3", "ansible_hostname": "managed-node3", "ansible_nodename": "managed-node3", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "ec25272c602494034078bc876e25857f", "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQC83yKiaGAYjfqsqlfpPMYFAYI2IZVpX8WgNfvPCdI/QOKjuzs4b6SbV/Bm0ogtp9Et9STTGoHBvp3tMYQ6i0y/9DHMBxtiHYJ+rIzJ/YNNMqKc+fMqexyOsi7rKKuzfxXrYU8uPBHq5WU3WAZmJxJn+UHZGog8hUnJ8momdJG+aYo9El3Qce4gVdwORcmHZUOa49M8lLCwTovtYArmkGETUVJ+Jk8huVTzYpASWxxcw6zOvUcn52HC6dmNQv/T+k2uW6UW0rybwIrVUlZXRNODrXs8kCGgOx1OI0XYB3FndJOnORF4A9Y6onLo/zUCEaO8Pi19mcfSbo2v+bmotTVk5jcmvR3jhVYJmJE6a+dQpjSZolSqMv8mI9tkztfxM6bJlNSZcTrvZEzu7cbiE38Pp/Ku143n9iGgWNmUQ2FhUDpoWEhoA767VUunE48P8ivpVZ/u5aEyupZSLEuWEvCLmmGPVcf9hVbcXw0n8RFvUwDdD8WARVhN5GJFUVN5JM0=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHm1sMDuVWGACN5pHFDkl6tR93F90YCY4cFGcXcCoQnN+oT963FmBwTMMlfDIm4G2OUATCZuz6QFZP9trAaUzXo=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIAB71QKijTWbanEvrb0ex0kLr0wX6qyv6naldRWNiIFP", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_local": {}, "ansible_apparmor": {"status": "disabled"}, "ansible_iscsi_iqn": "", "ansible_is_chroot": false, "ansible_loadavg": {"1m": 0.65869140625, "5m": 0.37109375, "15m": 0.17529296875}, "ansible_fips": false, "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "10", "ansible_distribution_major_version": "10", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_dns": {"search": ["us-east-1.a<<< 13731 1727203823.47623: stdout chunk (state=3): >>>ws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_lsb": {}, "ansible_python": {"version": {"major": 3, "minor": 12, "micro": 5, "releaselevel": "final", "serial": 0}, "version_info": [3, 12, 5, "final", 0], "executable": "/usr/bin/python3.12", "has_sslcontext": true, "type": "cpython"}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "50", "second": "23", "epoch": "1727203823", "epoch_int": "1727203823", "date": "2024-09-24", "time": "14:50:23", "iso8601_micro": "2024-09-24T18:50:23.112426Z", "iso8601": "2024-09-24T18:50:23Z", "iso8601_basic": "20240924T145023112426", "iso8601_basic_short": "20240924T145023", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_fibre_channel_wwn": [], "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_hostnqn": "nqn.2014-08.org.nvmexpress:uuid:11e86335-d786-4518-8abc-c9417b351256", "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3531, "ansible_memfree_mb": 2931, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3531, "used": 600, "free": 2931}, "nocache": {"free": 3267, "used": 264}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec25272c-6024-9403-4078-bc876e25857f", "ansible_product_uuid": "ec25272c-6024-9403-4078-bc876e25857f", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "512", "partitions": {"xvda2": {"links": {"ids": [], "uuids": ["973ca870-ed1b-4e56-a8b4-735608119a28"], "labels": [], "masters": []}, "start": "4096", "sectors": "524283871", "sectorsize": 512, "size": "250.00 GB", "uuid": "973ca870-ed1b-4e56-a8b4-735608119a28", "holders": []}, "xvda1": {"links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "start": "2048", "sectors": "2048", "sectorsize": 512, "size": "1.00 MB", "uuid": null, "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda2": ["973ca870-ed1b-4e56-a8b4-735608119a28"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 400, "ansible_lvm": {"lvs": {}, "vgs": {}, "pvs": {}}, "ansible_mounts": [{"mount": "/", "device": "/dev/xvda2", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "dump": 0, "passno": 0, "size_total": 268366229504, "size_available": 261796044800, "block_size": 4096, "block_total": 65519099, "block_available": 63915050, "block_used": 1604049, "inode_total": 131070960, "inode_available": 131027343, "inode_used": 43617, "uuid": "973ca870-ed1b-4e56-a8b4-735608119a28"}], "ansible_interfaces": ["lo", "eth0"], "ansible_eth0": {"device": "eth0", "macaddress": "02:83:38:1a:ae:4d", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.47.22", "broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0", "prefix": "22"}, "ipv6": [{"address": "fe80::83:38ff:fe1a:ae4d", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.44.1", "interface": "eth0", "address": "10.31.47.22", "broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0", "prefix": "22", "macaddress": "02:83:38:1a:ae:4d", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.47.22"], "ansible_all_ipv6_addresses": ["fe80::83:38ff:fe1a:ae4d"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.47.22", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::83:38ff:fe1a:ae4d"]}, "ansible_service_mgr": "systemd", "ansible_pkg_mgr": "dnf", "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} <<< 13731 1727203823.48567: stdout chunk (state=3): >>># clear sys.path_importer_cache # clear sys.path_hooks # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_exc # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanu<<< 13731 1727203823.48650: stdout chunk (state=3): >>>p[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing<<< 13731 1727203823.48709: stdout chunk (state=3): >>> ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _heapq # cleanup[2] removing heapq # destroy heapq # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # cleanup[2] removing multiprocessing.queues # cleanup[2] removing multiprocessing.synchronize # cleanup[2] removing multiprocessing.dummy.connection # cleanup[2] removing multiprocessing.dummy <<< 13731 1727203823.49099: stdout chunk (state=3): >>># destroy _sitebuiltins <<< 13731 1727203823.49145: stdout chunk (state=3): >>># destroy importlib.machinery # destroy importlib._abc # destroy importlib.util # destroy _bz2 # destroy _compression # destroy _lzma <<< 13731 1727203823.49187: stdout chunk (state=3): >>># destroy _blake2 # destroy binascii # destroy zlib # destroy bz2 # destroy lzma # destroy zipfile._path # destroy zipfile # destroy pathlib # destroy zipfile._path.glob # destroy ipaddress <<< 13731 1727203823.49346: stdout chunk (state=3): >>># destroy ntpath # destroy importlib # destroy zipimport # destroy __main__ # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy grp # destroy encodings # destroy _locale # destroy locale # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog # destroy uuid # destroy selinux # destroy shutil <<< 13731 1727203823.49349: stdout chunk (state=3): >>># destroy distro # destroy distro.distro # destroy argparse # destroy logging <<< 13731 1727203823.49391: stdout chunk (state=3): >>># destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector <<< 13731 1727203823.49427: stdout chunk (state=3): >>># destroy multiprocessing # destroy multiprocessing.queues # destroy multiprocessing.synchronize # destroy multiprocessing.dummy # destroy multiprocessing.pool # destroy signal # destroy pickle # destroy _compat_pickle # destroy _pickle <<< 13731 1727203823.49435: stdout chunk (state=3): >>># destroy queue # destroy _heapq # destroy _queue <<< 13731 1727203823.49469: stdout chunk (state=3): >>># destroy multiprocessing.reduction # destroy selectors # destroy shlex # destroy fcntl # destroy datetime <<< 13731 1727203823.49503: stdout chunk (state=3): >>># destroy subprocess # destroy base64 # destroy _ssl <<< 13731 1727203823.49531: stdout chunk (state=3): >>># destroy ansible.module_utils.compat.selinux <<< 13731 1727203823.49534: stdout chunk (state=3): >>># destroy getpass # destroy pwd <<< 13731 1727203823.49577: stdout chunk (state=3): >>># destroy termios # destroy json # destroy socket # destroy struct <<< 13731 1727203823.49581: stdout chunk (state=3): >>># destroy glob <<< 13731 1727203823.49655: stdout chunk (state=3): >>># destroy fnmatch # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # destroy unicodedata # destroy errno # destroy multiprocessing.connection <<< 13731 1727203823.49658: stdout chunk (state=3): >>># destroy tempfile # destroy multiprocessing.context # destroy multiprocessing.process # destroy multiprocessing.util # destroy _multiprocessing # destroy array # destroy multiprocessing.dummy.connection # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping configparser # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes <<< 13731 1727203823.49664: stdout chunk (state=3): >>># cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket <<< 13731 1727203823.49694: stdout chunk (state=3): >>># cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform <<< 13731 1727203823.49723: stdout chunk (state=3): >>># cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings <<< 13731 1727203823.49739: stdout chunk (state=3): >>># cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser <<< 13731 1727203823.49763: stdout chunk (state=3): >>># cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types <<< 13731 1727203823.49801: stdout chunk (state=3): >>># cleanup[3] wiping encodings.utf_8_sig # cleanup[3] wiping os # destroy posixpath # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime <<< 13731 1727203823.50034: stdout chunk (state=3): >>># destroy sys.monitoring # destroy _socket # destroy _collections # destroy platform <<< 13731 1727203823.50045: stdout chunk (state=3): >>># destroy _uuid # destroy stat # destroy genericpath # destroy re._parser # destroy tokenize <<< 13731 1727203823.50115: stdout chunk (state=3): >>># destroy ansible.module_utils.six.moves.urllib # destroy copyreg # destroy contextlib # destroy _typing <<< 13731 1727203823.50147: stdout chunk (state=3): >>># destroy _tokenize # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy operator # destroy ansible.module_utils.six.moves <<< 13731 1727203823.50158: stdout chunk (state=3): >>># destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal # clear sys.meta_path # clear sys.modules # destroy _frozen_importlib <<< 13731 1727203823.50322: stdout chunk (state=3): >>># destroy codecs # destroy encodings.aliases # destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy encodings.idna # destroy _codecs # destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections # destroy threading # destroy atexit # destroy _warnings # destroy math # destroy _bisect # destroy time <<< 13731 1727203823.50369: stdout chunk (state=3): >>># destroy _random # destroy _weakref # destroy _hashlib <<< 13731 1727203823.50486: stdout chunk (state=3): >>># destroy _operator # destroy _sre # destroy _string # destroy re # destroy itertools # destroy _abc # destroy posix # destroy _functools # destroy builtins # destroy _thread <<< 13731 1727203823.51202: stdout chunk (state=3): >>># clear sys.audit hooks <<< 13731 1727203823.51205: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203823.51208: stdout chunk (state=3): >>><<< 13731 1727203823.51214: stderr chunk (state=3): >>><<< 13731 1727203823.51730: _low_level_execute_command() done: rc=0, stdout=import _frozen_importlib # frozen import _imp # builtin import '_thread' # import '_warnings' # import '_weakref' # import '_io' # import 'marshal' # import 'posix' # import '_frozen_importlib_external' # # installing zipimport hook import 'time' # import 'zipimport' # # installed zipimport hook # /usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/encodings/__init__.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc' import '_codecs' # import 'codecs' # # /usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc matches /usr/lib64/python3.12/encodings/aliases.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79774184d0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79773e7b30> # /usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc' import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797741aa50> import '_signal' # import '_abc' # import 'abc' # import 'io' # import '_stat' # import 'stat' # import '_collections_abc' # import 'genericpath' # import 'posixpath' # import 'os' # import '_sitebuiltins' # Processing user site-packages Processing global site-packages Adding directory: '/usr/local/lib/python3.12/site-packages' Adding directory: '/usr/lib64/python3.12/site-packages' Adding directory: '/usr/lib/python3.12/site-packages' Processing .pth file: '/usr/lib/python3.12/site-packages/distutils-precedence.pth' # /usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8_sig.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc' import 'encodings.utf_8_sig' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977209130> # /usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/_distutils_hack/__init__.py # code object from '/usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc' import '_distutils_hack' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797720a060> import 'site' # Python 3.12.5 (main, Aug 23 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-1)] on linux Type "help", "copyright", "credits" or "license" for more information. # /usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc matches /usr/lib64/python3.12/base64.py # code object from '/usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc' # /usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/re/__init__.py # code object from '/usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc matches /usr/lib64/python3.12/enum.py # code object from '/usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/types.cpython-312.pyc matches /usr/lib64/python3.12/types.py # code object from '/usr/lib64/python3.12/__pycache__/types.cpython-312.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977247e90> # /usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc matches /usr/lib64/python3.12/operator.py # code object from '/usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc' import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977247f50> # /usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc matches /usr/lib64/python3.12/functools.py # code object from '/usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc' # /usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/collections/__init__.py # code object from '/usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc' import 'itertools' # # /usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc matches /usr/lib64/python3.12/keyword.py # code object from '/usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797727f890> # /usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc matches /usr/lib64/python3.12/reprlib.py # code object from '/usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797727ff20> import '_collections' # import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797725fb60> import '_functools' # import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797725d280> import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977245040> # /usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc matches /usr/lib64/python3.12/re/_compiler.py # code object from '/usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc' import '_sre' # # /usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc matches /usr/lib64/python3.12/re/_parser.py # code object from '/usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc' # /usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc matches /usr/lib64/python3.12/re/_constants.py # code object from '/usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc' import 're._constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772a3770> import 're._parser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772a2390> # /usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc matches /usr/lib64/python3.12/re/_casefix.py # code object from '/usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc' import 're._casefix' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797725e120> import 're._compiler' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977246900> # /usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc matches /usr/lib64/python3.12/copyreg.py # code object from '/usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d4830> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772442c0> # /usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc matches /usr/lib64/python3.12/struct.py # code object from '/usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772d4ce0> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d4b90> # extension module 'binascii' loaded from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772d4f80> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977242de0> # /usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/importlib/__init__.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc matches /usr/lib64/python3.12/warnings.py # code object from '/usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d5640> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d5310> import 'importlib.machinery' # # /usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc matches /usr/lib64/python3.12/importlib/_abc.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc' import 'importlib._abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d6510> import 'importlib.util' # import 'runpy' # # /usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc matches /usr/lib64/python3.12/shutil.py # code object from '/usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc matches /usr/lib64/python3.12/fnmatch.py # code object from '/usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772ec710> import 'errno' # # extension module 'zlib' loaded from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772eddc0> # /usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc matches /usr/lib64/python3.12/bz2.py # code object from '/usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc matches /usr/lib64/python3.12/_compression.py # code object from '/usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772eec60> # extension module '_bz2' loaded from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772ef290> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772ee1b0> # /usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc matches /usr/lib64/python3.12/lzma.py # code object from '/usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc' # extension module '_lzma' loaded from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79772efd10> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772ef440> import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d6480> # /usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc matches /usr/lib64/python3.12/tempfile.py # code object from '/usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/random.cpython-312.pyc matches /usr/lib64/python3.12/random.py # code object from '/usr/lib64/python3.12/__pycache__/random.cpython-312.pyc' # extension module 'math' loaded from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976ff7c80> # /usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc matches /usr/lib64/python3.12/bisect.py # code object from '/usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc' # extension module '_bisect' loaded from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79770206b0> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977020440> # extension module '_random' loaded from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79770206e0> # /usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc matches /usr/lib64/python3.12/hashlib.py # code object from '/usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc' # extension module '_hashlib' loaded from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7977021010> # extension module '_blake2' loaded from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7977021970> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79770208c0> import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976ff5e20> # /usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc matches /usr/lib64/python3.12/weakref.py # code object from '/usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc matches /usr/lib64/python3.12/_weakrefset.py # code object from '/usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977022d20> import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977020e90> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79772d6c30> # /usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/__init__.py # code object from '/usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc matches /usr/lib64/python3.12/threading.py # code object from '/usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc' import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797704f080> # /usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/__init__.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc matches /usr/lib64/python3.12/contextlib.py # code object from '/usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc' import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977073410> # /usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc matches /usr/lib64/python3.12/pathlib.py # code object from '/usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc' import 'ntpath' # # /usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/urllib/__init__.py # code object from '/usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc' import 'urllib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79770d01d0> # /usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc matches /usr/lib64/python3.12/urllib/parse.py # code object from '/usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc matches /usr/lib64/python3.12/ipaddress.py # code object from '/usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc' import 'ipaddress' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79770d2930> import 'urllib.parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79770d02f0> import 'pathlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797709d1f0> # /usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/glob.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc' import 'zipfile._path.glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797709d9a0> import 'zipfile._path' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977072210> import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7977023c50> # code object from '/usr/lib64/python3.12/encodings/cp437.pyc' import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f7977072330> # zipimport: found 103 names in '/tmp/ansible_ansible.legacy.setup_payload_r1p_k8g4/ansible_ansible.legacy.setup_payload.zip' # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc matches /usr/lib64/python3.12/pkgutil.py # code object from '/usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc matches /usr/lib64/python3.12/typing.py # code object from '/usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc' # /usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc matches /usr/lib64/python3.12/collections/abc.py # code object from '/usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797698aff0> import '_typing' # import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976969ee0> import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976969070> # zipimport: zlib available import 'ansible' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc matches /usr/lib64/python3.12/__future__.py # code object from '/usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976988e90> # /usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/json/__init__.py # code object from '/usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc matches /usr/lib64/python3.12/json/decoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc' # /usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc matches /usr/lib64/python3.12/json/scanner.py # code object from '/usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc' # extension module '_json' loaded from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79769c2960> import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79769c26f0> import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79769c2000> # /usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc matches /usr/lib64/python3.12/json/encoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc' import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79769c2a50> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797698ba10> import 'atexit' # # extension module 'grp' loaded from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79769c36e0> # extension module 'fcntl' loaded from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79769c3920> # /usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc matches /usr/lib64/python3.12/locale.py # code object from '/usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc' import '_locale' # import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79769c3e30> import 'pwd' # # /usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc matches /usr/lib64/python3.12/platform.py # code object from '/usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc' import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976825b20> # extension module 'select' loaded from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79768277d0> # /usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc matches /usr/lib64/python3.12/selectors.py # code object from '/usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc' import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682c1a0> # /usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc matches /usr/lib64/python3.12/shlex.py # code object from '/usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc' import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682d340> # /usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc matches /usr/lib64/python3.12/subprocess.py # code object from '/usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc matches /usr/lib64/python3.12/signal.py # code object from '/usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc' import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682fda0> # extension module '_posixsubprocess' loaded from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797682fe30> import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682e090> # /usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc matches /usr/lib64/python3.12/traceback.py # code object from '/usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc matches /usr/lib64/python3.12/linecache.py # code object from '/usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc matches /usr/lib64/python3.12/tokenize.py # code object from '/usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/token.cpython-312.pyc matches /usr/lib64/python3.12/token.py # code object from '/usr/lib64/python3.12/__pycache__/token.cpython-312.pyc' import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976833c80> import '_tokenize' # import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976832780> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768324e0> # /usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc matches /usr/lib64/python3.12/textwrap.py # code object from '/usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc' import 'textwrap' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976832a50> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797682e570> # extension module 'syslog' loaded from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976877f80> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768784a0> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/journal.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc matches /usr/lib64/python3.12/datetime.py # code object from '/usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc' # extension module '_datetime' loaded from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976879af0> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768798b0> # /usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc matches /usr/lib64/python3.12/uuid.py # code object from '/usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc' # extension module '_uuid' loaded from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797687bfe0> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797687a1b0> # /usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/logging/__init__.py # code object from '/usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/string.cpython-312.pyc matches /usr/lib64/python3.12/string.py # code object from '/usr/lib64/python3.12/__pycache__/string.cpython-312.pyc' import '_string' # import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797687f830> import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797687c200> # extension module 'systemd._journal' loaded from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79768805f0> # extension module 'systemd._reader' loaded from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._reader' executed from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976880650> # extension module 'systemd.id128' loaded from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976880bf0> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976878200> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc matches /usr/lib64/python3.12/socket.py # code object from '/usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc' # extension module '_socket' loaded from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797670c350> # extension module 'array' loaded from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797670d670> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976882ae0> # extension module 'systemd._daemon' loaded from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976883e60> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768826c0> # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.text' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.six' # import 'ansible.module_utils.six.moves' # import 'ansible.module_utils.six.moves.collections_abc' # import 'ansible.module_utils.common.text.converters' # # /usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/__init__.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc' # extension module '_ctypes' loaded from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976711760> # /usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/_endian.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc' import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976712510> import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797670d850> import 'ansible.module_utils.compat.selinux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils._text' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc matches /usr/lib64/python3.12/copy.py # code object from '/usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc' import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767124b0> # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.collections' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.warnings' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.errors' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.parsing' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.parsing.convert_bool' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc matches /usr/lib64/python3.12/ast.py # code object from '/usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc' import '_ast' # import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976713680> # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.text.formatters' # import 'ansible.module_utils.common.validation' # import 'ansible.module_utils.common.parameters' # import 'ansible.module_utils.common.arg_spec' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.locale' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/selinux/__init__.py # code object from '/usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc' # extension module 'selinux._selinux' loaded from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f797671e030> import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976719040> import 'ansible.module_utils.common.file' # import 'ansible.module_utils.common.process' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/__init__.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc' # /usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/distro.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc matches /usr/lib64/python3.12/argparse.py # code object from '/usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc matches /usr/lib64/python3.12/gettext.py # code object from '/usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc' import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976806ae0> import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79768fe7b0> import 'distro.distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797671e270> import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976714890> # destroy ansible.module_utils.distro import 'ansible.module_utils.distro' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common._utils' # import 'ansible.module_utils.common.sys_info' # import 'ansible.module_utils.basic' # # zipimport: zlib available # zipimport: zlib available import 'ansible.modules' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.namespace' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat.typing' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/multiprocessing/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/__init__.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/multiprocessing/__pycache__/context.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/context.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/context.cpython-312.pyc' # /usr/lib64/python3.12/multiprocessing/__pycache__/process.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/process.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/process.cpython-312.pyc' import 'multiprocessing.process' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b23f0> # /usr/lib64/python3.12/multiprocessing/__pycache__/reduction.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/reduction.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/reduction.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/pickle.cpython-312.pyc matches /usr/lib64/python3.12/pickle.py # code object from '/usr/lib64/python3.12/__pycache__/pickle.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_compat_pickle.cpython-312.pyc matches /usr/lib64/python3.12/_compat_pickle.py # code object from '/usr/lib64/python3.12/__pycache__/_compat_pickle.cpython-312.pyc' import '_compat_pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976420110> # extension module '_pickle' loaded from '/usr/lib64/python3.12/lib-dynload/_pickle.cpython-312-x86_64-linux-gnu.so' # extension module '_pickle' executed from '/usr/lib64/python3.12/lib-dynload/_pickle.cpython-312-x86_64-linux-gnu.so' import '_pickle' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79764206e0> import 'pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f797679f050> import 'multiprocessing.reduction' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b2f90> import 'multiprocessing.context' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b0aa0> import 'multiprocessing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b13a0> # /usr/lib64/python3.12/multiprocessing/__pycache__/pool.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/pool.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/pool.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/queue.cpython-312.pyc matches /usr/lib64/python3.12/queue.py # code object from '/usr/lib64/python3.12/__pycache__/queue.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/heapq.cpython-312.pyc matches /usr/lib64/python3.12/heapq.py # code object from '/usr/lib64/python3.12/__pycache__/heapq.cpython-312.pyc' # extension module '_heapq' loaded from '/usr/lib64/python3.12/lib-dynload/_heapq.cpython-312-x86_64-linux-gnu.so' # extension module '_heapq' executed from '/usr/lib64/python3.12/lib-dynload/_heapq.cpython-312-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79764233e0> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976422c90> # extension module '_queue' loaded from '/usr/lib64/python3.12/lib-dynload/_queue.cpython-312-x86_64-linux-gnu.so' # extension module '_queue' executed from '/usr/lib64/python3.12/lib-dynload/_queue.cpython-312-x86_64-linux-gnu.so' import '_queue' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976422e70> import 'queue' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79764220f0> # /usr/lib64/python3.12/multiprocessing/__pycache__/util.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/util.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/util.cpython-312.pyc' import 'multiprocessing.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976423530> # /usr/lib64/python3.12/multiprocessing/__pycache__/connection.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/connection.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/connection.cpython-312.pyc' # extension module '_multiprocessing' loaded from '/usr/lib64/python3.12/lib-dynload/_multiprocessing.cpython-312-x86_64-linux-gnu.so' # extension module '_multiprocessing' executed from '/usr/lib64/python3.12/lib-dynload/_multiprocessing.cpython-312-x86_64-linux-gnu.so' import '_multiprocessing' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976486060> import 'multiprocessing.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976423f80> import 'multiprocessing.pool' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79767b1550> import 'ansible.module_utils.facts.timeout' # import 'ansible.module_utils.facts.collector' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.other' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.other.facter' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.other.ohai' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.apparmor' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.caps' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.chroot' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.utils' # import 'ansible.module_utils.facts.system.cmdline' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.distribution' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat.datetime' # import 'ansible.module_utils.facts.system.date_time' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.env' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.dns' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.fips' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.loadavg' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/glob.py # code object from '/usr/lib64/python3.12/__pycache__/glob.cpython-312.pyc' import 'glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976487830> # /usr/lib64/python3.12/__pycache__/configparser.cpython-312.pyc matches /usr/lib64/python3.12/configparser.py # code object from '/usr/lib64/python3.12/__pycache__/configparser.cpython-312.pyc' import 'configparser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976486bd0> import 'ansible.module_utils.facts.system.local' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.lsb' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.pkg_mgr' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.platform' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/ssl.cpython-312.pyc matches /usr/lib64/python3.12/ssl.py # code object from '/usr/lib64/python3.12/__pycache__/ssl.cpython-312.pyc' # extension module '_ssl' loaded from '/usr/lib64/python3.12/lib-dynload/_ssl.cpython-312-x86_64-linux-gnu.so' # extension module '_ssl' executed from '/usr/lib64/python3.12/lib-dynload/_ssl.cpython-312-x86_64-linux-gnu.so' import '_ssl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79764be270> import 'ssl' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79764af140> import 'ansible.module_utils.facts.system.python' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.selinux' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat.version' # import 'ansible.module_utils.facts.system.service_mgr' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.ssh_pub_keys' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/getpass.cpython-312.pyc matches /usr/lib64/python3.12/getpass.py # code object from '/usr/lib64/python3.12/__pycache__/getpass.cpython-312.pyc' # extension module 'termios' loaded from '/usr/lib64/python3.12/lib-dynload/termios.cpython-312-x86_64-linux-gnu.so' # extension module 'termios' executed from '/usr/lib64/python3.12/lib-dynload/termios.cpython-312-x86_64-linux-gnu.so' import 'termios' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f79764d1dc0> import 'getpass' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79764be060> import 'ansible.module_utils.facts.system.user' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.base' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.aix' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.sysctl' # import 'ansible.module_utils.facts.hardware.darwin' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.freebsd' # import 'ansible.module_utils.facts.hardware.dragonfly' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.hpux' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.linux' # import 'ansible.module_utils.facts.hardware.hurd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.netbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.openbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.sunos' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.base' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.generic_bsd' # import 'ansible.module_utils.facts.network.aix' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.darwin' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.dragonfly' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.fc_wwn' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.freebsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.hpux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.hurd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.linux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.iscsi' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.nvme' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.netbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.openbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.sunos' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.base' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.sysctl' # import 'ansible.module_utils.facts.virtual.freebsd' # import 'ansible.module_utils.facts.virtual.dragonfly' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.hpux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.linux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.netbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.openbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.sunos' # import 'ansible.module_utils.facts.default_collectors' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.ansible_collector' # import 'ansible.module_utils.facts.compat' # import 'ansible.module_utils.facts' # # zipimport: zlib available # /usr/lib64/python3.12/encodings/__pycache__/idna.cpython-312.pyc matches /usr/lib64/python3.12/encodings/idna.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/idna.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/stringprep.cpython-312.pyc matches /usr/lib64/python3.12/stringprep.py # code object from '/usr/lib64/python3.12/__pycache__/stringprep.cpython-312.pyc' # extension module 'unicodedata' loaded from '/usr/lib64/python3.12/lib-dynload/unicodedata.cpython-312-x86_64-linux-gnu.so' # extension module 'unicodedata' executed from '/usr/lib64/python3.12/lib-dynload/unicodedata.cpython-312-x86_64-linux-gnu.so' import 'unicodedata' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f7976273860> import 'stringprep' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976270740> import 'encodings.idna' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976273140> # /usr/lib64/python3.12/multiprocessing/__pycache__/queues.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/queues.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/queues.cpython-312.pyc' import 'multiprocessing.queues' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79762b87a0> # /usr/lib64/python3.12/multiprocessing/__pycache__/synchronize.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/synchronize.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/synchronize.cpython-312.pyc' import 'multiprocessing.synchronize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79762b96a0> # /usr/lib64/python3.12/multiprocessing/dummy/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/dummy/__init__.py # code object from '/usr/lib64/python3.12/multiprocessing/dummy/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/multiprocessing/dummy/__pycache__/connection.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/dummy/connection.py # code object from '/usr/lib64/python3.12/multiprocessing/dummy/__pycache__/connection.cpython-312.pyc' import 'multiprocessing.dummy.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f7976306420> import 'multiprocessing.dummy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f79762bb4d0> PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame PyThreadState_Clear: warning: thread still has a frame {"ansible_facts": {"ansible_env": {"PYTHONVERBOSE": "1", "SHELL": "/bin/bash", "GPG_TTY": "/dev/pts/0", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.12", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.45.138 53526 10.31.47.22 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.45.138 53526 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0"}, "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "Super User", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_system": "Linux", "ansible_kernel": "6.11.0-25.el10.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Mon Sep 16 20:35:26 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.12.5", "ansible_fqdn": "managed-node3", "ansible_hostname": "managed-node3", "ansible_nodename": "managed-node3", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "ec25272c602494034078bc876e25857f", "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQC83yKiaGAYjfqsqlfpPMYFAYI2IZVpX8WgNfvPCdI/QOKjuzs4b6SbV/Bm0ogtp9Et9STTGoHBvp3tMYQ6i0y/9DHMBxtiHYJ+rIzJ/YNNMqKc+fMqexyOsi7rKKuzfxXrYU8uPBHq5WU3WAZmJxJn+UHZGog8hUnJ8momdJG+aYo9El3Qce4gVdwORcmHZUOa49M8lLCwTovtYArmkGETUVJ+Jk8huVTzYpASWxxcw6zOvUcn52HC6dmNQv/T+k2uW6UW0rybwIrVUlZXRNODrXs8kCGgOx1OI0XYB3FndJOnORF4A9Y6onLo/zUCEaO8Pi19mcfSbo2v+bmotTVk5jcmvR3jhVYJmJE6a+dQpjSZolSqMv8mI9tkztfxM6bJlNSZcTrvZEzu7cbiE38Pp/Ku143n9iGgWNmUQ2FhUDpoWEhoA767VUunE48P8ivpVZ/u5aEyupZSLEuWEvCLmmGPVcf9hVbcXw0n8RFvUwDdD8WARVhN5GJFUVN5JM0=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHm1sMDuVWGACN5pHFDkl6tR93F90YCY4cFGcXcCoQnN+oT963FmBwTMMlfDIm4G2OUATCZuz6QFZP9trAaUzXo=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIAB71QKijTWbanEvrb0ex0kLr0wX6qyv6naldRWNiIFP", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_local": {}, "ansible_apparmor": {"status": "disabled"}, "ansible_iscsi_iqn": "", "ansible_is_chroot": false, "ansible_loadavg": {"1m": 0.65869140625, "5m": 0.37109375, "15m": 0.17529296875}, "ansible_fips": false, "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "10", "ansible_distribution_major_version": "10", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_lsb": {}, "ansible_python": {"version": {"major": 3, "minor": 12, "micro": 5, "releaselevel": "final", "serial": 0}, "version_info": [3, 12, 5, "final", 0], "executable": "/usr/bin/python3.12", "has_sslcontext": true, "type": "cpython"}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "50", "second": "23", "epoch": "1727203823", "epoch_int": "1727203823", "date": "2024-09-24", "time": "14:50:23", "iso8601_micro": "2024-09-24T18:50:23.112426Z", "iso8601": "2024-09-24T18:50:23Z", "iso8601_basic": "20240924T145023112426", "iso8601_basic_short": "20240924T145023", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_fibre_channel_wwn": [], "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_hostnqn": "nqn.2014-08.org.nvmexpress:uuid:11e86335-d786-4518-8abc-c9417b351256", "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3531, "ansible_memfree_mb": 2931, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3531, "used": 600, "free": 2931}, "nocache": {"free": 3267, "used": 264}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec25272c-6024-9403-4078-bc876e25857f", "ansible_product_uuid": "ec25272c-6024-9403-4078-bc876e25857f", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "512", "partitions": {"xvda2": {"links": {"ids": [], "uuids": ["973ca870-ed1b-4e56-a8b4-735608119a28"], "labels": [], "masters": []}, "start": "4096", "sectors": "524283871", "sectorsize": 512, "size": "250.00 GB", "uuid": "973ca870-ed1b-4e56-a8b4-735608119a28", "holders": []}, "xvda1": {"links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "start": "2048", "sectors": "2048", "sectorsize": 512, "size": "1.00 MB", "uuid": null, "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda2": ["973ca870-ed1b-4e56-a8b4-735608119a28"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 400, "ansible_lvm": {"lvs": {}, "vgs": {}, "pvs": {}}, "ansible_mounts": [{"mount": "/", "device": "/dev/xvda2", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "dump": 0, "passno": 0, "size_total": 268366229504, "size_available": 261796044800, "block_size": 4096, "block_total": 65519099, "block_available": 63915050, "block_used": 1604049, "inode_total": 131070960, "inode_available": 131027343, "inode_used": 43617, "uuid": "973ca870-ed1b-4e56-a8b4-735608119a28"}], "ansible_interfaces": ["lo", "eth0"], "ansible_eth0": {"device": "eth0", "macaddress": "02:83:38:1a:ae:4d", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.47.22", "broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0", "prefix": "22"}, "ipv6": [{"address": "fe80::83:38ff:fe1a:ae4d", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.44.1", "interface": "eth0", "address": "10.31.47.22", "broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0", "prefix": "22", "macaddress": "02:83:38:1a:ae:4d", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.47.22"], "ansible_all_ipv6_addresses": ["fe80::83:38ff:fe1a:ae4d"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.47.22", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::83:38ff:fe1a:ae4d"]}, "ansible_service_mgr": "systemd", "ansible_pkg_mgr": "dnf", "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} # clear sys.path_importer_cache # clear sys.path_hooks # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_exc # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _heapq # cleanup[2] removing heapq # destroy heapq # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # cleanup[2] removing multiprocessing.queues # cleanup[2] removing multiprocessing.synchronize # cleanup[2] removing multiprocessing.dummy.connection # cleanup[2] removing multiprocessing.dummy # destroy _sitebuiltins # destroy importlib.machinery # destroy importlib._abc # destroy importlib.util # destroy _bz2 # destroy _compression # destroy _lzma # destroy _blake2 # destroy binascii # destroy zlib # destroy bz2 # destroy lzma # destroy zipfile._path # destroy zipfile # destroy pathlib # destroy zipfile._path.glob # destroy ipaddress # destroy ntpath # destroy importlib # destroy zipimport # destroy __main__ # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy grp # destroy encodings # destroy _locale # destroy locale # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog # destroy uuid # destroy selinux # destroy shutil # destroy distro # destroy distro.distro # destroy argparse # destroy logging # destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.queues # destroy multiprocessing.synchronize # destroy multiprocessing.dummy # destroy multiprocessing.pool # destroy signal # destroy pickle # destroy _compat_pickle # destroy _pickle # destroy queue # destroy _heapq # destroy _queue # destroy multiprocessing.reduction # destroy selectors # destroy shlex # destroy fcntl # destroy datetime # destroy subprocess # destroy base64 # destroy _ssl # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy pwd # destroy termios # destroy json # destroy socket # destroy struct # destroy glob # destroy fnmatch # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # destroy unicodedata # destroy errno # destroy multiprocessing.connection # destroy tempfile # destroy multiprocessing.context # destroy multiprocessing.process # destroy multiprocessing.util # destroy _multiprocessing # destroy array # destroy multiprocessing.dummy.connection # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping configparser # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform # cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser # cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types # cleanup[3] wiping encodings.utf_8_sig # cleanup[3] wiping os # destroy posixpath # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy sys.monitoring # destroy _socket # destroy _collections # destroy platform # destroy _uuid # destroy stat # destroy genericpath # destroy re._parser # destroy tokenize # destroy ansible.module_utils.six.moves.urllib # destroy copyreg # destroy contextlib # destroy _typing # destroy _tokenize # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy operator # destroy ansible.module_utils.six.moves # destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal # clear sys.meta_path # clear sys.modules # destroy _frozen_importlib # destroy codecs # destroy encodings.aliases # destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy encodings.idna # destroy _codecs # destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections # destroy threading # destroy atexit # destroy _warnings # destroy math # destroy _bisect # destroy time # destroy _random # destroy _weakref # destroy _hashlib # destroy _operator # destroy _sre # destroy _string # destroy re # destroy itertools # destroy _abc # destroy posix # destroy _functools # destroy builtins # destroy _thread # clear sys.audit hooks , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. [WARNING]: Module invocation had junk after the JSON data: # clear sys.path_importer_cache # clear sys.path_hooks # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_exc # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _heapq # cleanup[2] removing heapq # destroy heapq # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # cleanup[2] removing multiprocessing.queues # cleanup[2] removing multiprocessing.synchronize # cleanup[2] removing multiprocessing.dummy.connection # cleanup[2] removing multiprocessing.dummy # destroy _sitebuiltins # destroy importlib.machinery # destroy importlib._abc # destroy importlib.util # destroy _bz2 # destroy _compression # destroy _lzma # destroy _blake2 # destroy binascii # destroy zlib # destroy bz2 # destroy lzma # destroy zipfile._path # destroy zipfile # destroy pathlib # destroy zipfile._path.glob # destroy ipaddress # destroy ntpath # destroy importlib # destroy zipimport # destroy __main__ # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy grp # destroy encodings # destroy _locale # destroy locale # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog # destroy uuid # destroy selinux # destroy shutil # destroy distro # destroy distro.distro # destroy argparse # destroy logging # destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.queues # destroy multiprocessing.synchronize # destroy multiprocessing.dummy # destroy multiprocessing.pool # destroy signal # destroy pickle # destroy _compat_pickle # destroy _pickle # destroy queue # destroy _heapq # destroy _queue # destroy multiprocessing.reduction # destroy selectors # destroy shlex # destroy fcntl # destroy datetime # destroy subprocess # destroy base64 # destroy _ssl # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy pwd # destroy termios # destroy json # destroy socket # destroy struct # destroy glob # destroy fnmatch # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # destroy unicodedata # destroy errno # destroy multiprocessing.connection # destroy tempfile # destroy multiprocessing.context # destroy multiprocessing.process # destroy multiprocessing.util # destroy _multiprocessing # destroy array # destroy multiprocessing.dummy.connection # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping configparser # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform # cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser # cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types # cleanup[3] wiping encodings.utf_8_sig # cleanup[3] wiping os # destroy posixpath # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy sys.monitoring # destroy _socket # destroy _collections # destroy platform # destroy _uuid # destroy stat # destroy genericpath # destroy re._parser # destroy tokenize # destroy ansible.module_utils.six.moves.urllib # destroy copyreg # destroy contextlib # destroy _typing # destroy _tokenize # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy operator # destroy ansible.module_utils.six.moves # destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal # clear sys.meta_path # clear sys.modules # destroy _frozen_importlib # destroy codecs # destroy encodings.aliases # destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy encodings.idna # destroy _codecs # destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections # destroy threading # destroy atexit # destroy _warnings # destroy math # destroy _bisect # destroy time # destroy _random # destroy _weakref # destroy _hashlib # destroy _operator # destroy _sre # destroy _string # destroy re # destroy itertools # destroy _abc # destroy posix # destroy _functools # destroy builtins # destroy _thread # clear sys.audit hooks [WARNING]: Platform linux on host managed-node3 is using the discovered Python interpreter at /usr/bin/python3.12, but future installation of another Python interpreter could change the meaning of that path. See https://docs.ansible.com/ansible- core/2.17/reference_appendices/interpreter_discovery.html for more information. 13731 1727203823.53995: done with _execute_module (ansible.legacy.setup, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.setup', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203823.53998: _low_level_execute_command(): starting 13731 1727203823.54000: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203821.860213-13800-179669116395604/ > /dev/null 2>&1 && sleep 0' 13731 1727203823.54792: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203823.54804: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203823.54880: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203823.57440: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203823.57580: stderr chunk (state=3): >>><<< 13731 1727203823.57583: stdout chunk (state=3): >>><<< 13731 1727203823.57586: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203823.57592: handler run complete 13731 1727203823.57654: variable 'ansible_facts' from source: unknown 13731 1727203823.57781: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203823.58070: variable 'ansible_facts' from source: unknown 13731 1727203823.58148: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203823.58278: attempt loop complete, returning result 13731 1727203823.58281: _execute() done 13731 1727203823.58284: dumping result to json 13731 1727203823.58332: done dumping result, returning 13731 1727203823.58335: done running TaskExecutor() for managed-node3/TASK: Gathering Facts [028d2410-947f-82dc-c122-000000000015] 13731 1727203823.58338: sending task result for task 028d2410-947f-82dc-c122-000000000015 13731 1727203823.58764: done sending task result for task 028d2410-947f-82dc-c122-000000000015 13731 1727203823.58767: WORKER PROCESS EXITING ok: [managed-node3] 13731 1727203823.59137: no more pending results, returning what we have 13731 1727203823.59140: results queue empty 13731 1727203823.59141: checking for any_errors_fatal 13731 1727203823.59142: done checking for any_errors_fatal 13731 1727203823.59143: checking for max_fail_percentage 13731 1727203823.59144: done checking for max_fail_percentage 13731 1727203823.59145: checking to see if all hosts have failed and the running result is not ok 13731 1727203823.59145: done checking to see if all hosts have failed 13731 1727203823.59146: getting the remaining hosts for this loop 13731 1727203823.59148: done getting the remaining hosts for this loop 13731 1727203823.59152: getting the next task for host managed-node3 13731 1727203823.59158: done getting next task for host managed-node3 13731 1727203823.59162: ^ task is: TASK: meta (flush_handlers) 13731 1727203823.59164: ^ state is: HOST STATE: block=1, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203823.59168: getting variables 13731 1727203823.59170: in VariableManager get_vars() 13731 1727203823.59193: Calling all_inventory to load vars for managed-node3 13731 1727203823.59196: Calling groups_inventory to load vars for managed-node3 13731 1727203823.59199: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203823.59208: Calling all_plugins_play to load vars for managed-node3 13731 1727203823.59210: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203823.59213: Calling groups_plugins_play to load vars for managed-node3 13731 1727203823.59399: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203823.59595: done with get_vars() 13731 1727203823.59606: done getting variables 13731 1727203823.59666: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__ 13731 1727203823.59725: in VariableManager get_vars() 13731 1727203823.59734: Calling all_inventory to load vars for managed-node3 13731 1727203823.59737: Calling groups_inventory to load vars for managed-node3 13731 1727203823.59739: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203823.59744: Calling all_plugins_play to load vars for managed-node3 13731 1727203823.59746: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203823.59749: Calling groups_plugins_play to load vars for managed-node3 13731 1727203823.60104: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203823.60390: done with get_vars() 13731 1727203823.60404: done queuing things up, now waiting for results queue to drain 13731 1727203823.60406: results queue empty 13731 1727203823.60407: checking for any_errors_fatal 13731 1727203823.60410: done checking for any_errors_fatal 13731 1727203823.60410: checking for max_fail_percentage 13731 1727203823.60412: done checking for max_fail_percentage 13731 1727203823.60412: checking to see if all hosts have failed and the running result is not ok 13731 1727203823.60413: done checking to see if all hosts have failed 13731 1727203823.60418: getting the remaining hosts for this loop 13731 1727203823.60420: done getting the remaining hosts for this loop 13731 1727203823.60422: getting the next task for host managed-node3 13731 1727203823.60427: done getting next task for host managed-node3 13731 1727203823.60429: ^ task is: TASK: Include the task 'el_repo_setup.yml' 13731 1727203823.60431: ^ state is: HOST STATE: block=2, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203823.60433: getting variables 13731 1727203823.60434: in VariableManager get_vars() 13731 1727203823.60442: Calling all_inventory to load vars for managed-node3 13731 1727203823.60444: Calling groups_inventory to load vars for managed-node3 13731 1727203823.60446: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203823.60451: Calling all_plugins_play to load vars for managed-node3 13731 1727203823.60453: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203823.60456: Calling groups_plugins_play to load vars for managed-node3 13731 1727203823.60833: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203823.61232: done with get_vars() 13731 1727203823.61240: done getting variables TASK [Include the task 'el_repo_setup.yml'] ************************************ task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_options_nm.yml:11 Tuesday 24 September 2024 14:50:23 -0400 (0:00:01.799) 0:00:01.844 ***** 13731 1727203823.61343: entering _queue_task() for managed-node3/include_tasks 13731 1727203823.61345: Creating lock for include_tasks 13731 1727203823.61677: worker is 1 (out of 1 available) 13731 1727203823.61691: exiting _queue_task() for managed-node3/include_tasks 13731 1727203823.61703: done queuing things up, now waiting for results queue to drain 13731 1727203823.61705: waiting for pending results... 13731 1727203823.62094: running TaskExecutor() for managed-node3/TASK: Include the task 'el_repo_setup.yml' 13731 1727203823.62104: in run() - task 028d2410-947f-82dc-c122-000000000006 13731 1727203823.62107: variable 'ansible_search_path' from source: unknown 13731 1727203823.62110: calling self._execute() 13731 1727203823.62144: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203823.62149: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203823.62159: variable 'omit' from source: magic vars 13731 1727203823.62263: _execute() done 13731 1727203823.62267: dumping result to json 13731 1727203823.62270: done dumping result, returning 13731 1727203823.62294: done running TaskExecutor() for managed-node3/TASK: Include the task 'el_repo_setup.yml' [028d2410-947f-82dc-c122-000000000006] 13731 1727203823.62298: sending task result for task 028d2410-947f-82dc-c122-000000000006 13731 1727203823.62447: done sending task result for task 028d2410-947f-82dc-c122-000000000006 13731 1727203823.62451: WORKER PROCESS EXITING 13731 1727203823.62495: no more pending results, returning what we have 13731 1727203823.62500: in VariableManager get_vars() 13731 1727203823.62531: Calling all_inventory to load vars for managed-node3 13731 1727203823.62534: Calling groups_inventory to load vars for managed-node3 13731 1727203823.62538: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203823.62549: Calling all_plugins_play to load vars for managed-node3 13731 1727203823.62552: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203823.62555: Calling groups_plugins_play to load vars for managed-node3 13731 1727203823.62810: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203823.62994: done with get_vars() 13731 1727203823.63001: variable 'ansible_search_path' from source: unknown 13731 1727203823.63013: we have included files to process 13731 1727203823.63014: generating all_blocks data 13731 1727203823.63015: done generating all_blocks data 13731 1727203823.63016: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml 13731 1727203823.63017: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml 13731 1727203823.63019: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml 13731 1727203823.63931: in VariableManager get_vars() 13731 1727203823.63946: done with get_vars() 13731 1727203823.63956: done processing included file 13731 1727203823.63958: iterating over new_blocks loaded from include file 13731 1727203823.63962: in VariableManager get_vars() 13731 1727203823.63972: done with get_vars() 13731 1727203823.63973: filtering new block on tags 13731 1727203823.64091: done filtering new block on tags 13731 1727203823.64095: in VariableManager get_vars() 13731 1727203823.64105: done with get_vars() 13731 1727203823.64107: filtering new block on tags 13731 1727203823.64122: done filtering new block on tags 13731 1727203823.64125: in VariableManager get_vars() 13731 1727203823.64135: done with get_vars() 13731 1727203823.64137: filtering new block on tags 13731 1727203823.64150: done filtering new block on tags 13731 1727203823.64151: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml for managed-node3 13731 1727203823.64157: extending task lists for all hosts with included blocks 13731 1727203823.64311: done extending task lists 13731 1727203823.64313: done processing included files 13731 1727203823.64314: results queue empty 13731 1727203823.64314: checking for any_errors_fatal 13731 1727203823.64315: done checking for any_errors_fatal 13731 1727203823.64316: checking for max_fail_percentage 13731 1727203823.64317: done checking for max_fail_percentage 13731 1727203823.64318: checking to see if all hosts have failed and the running result is not ok 13731 1727203823.64318: done checking to see if all hosts have failed 13731 1727203823.64319: getting the remaining hosts for this loop 13731 1727203823.64320: done getting the remaining hosts for this loop 13731 1727203823.64322: getting the next task for host managed-node3 13731 1727203823.64326: done getting next task for host managed-node3 13731 1727203823.64328: ^ task is: TASK: Gather the minimum subset of ansible_facts required by the network role test 13731 1727203823.64330: ^ state is: HOST STATE: block=2, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203823.64332: getting variables 13731 1727203823.64333: in VariableManager get_vars() 13731 1727203823.64341: Calling all_inventory to load vars for managed-node3 13731 1727203823.64343: Calling groups_inventory to load vars for managed-node3 13731 1727203823.64345: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203823.64351: Calling all_plugins_play to load vars for managed-node3 13731 1727203823.64354: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203823.64356: Calling groups_plugins_play to load vars for managed-node3 13731 1727203823.64716: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203823.65107: done with get_vars() 13731 1727203823.65115: done getting variables TASK [Gather the minimum subset of ansible_facts required by the network role test] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:3 Tuesday 24 September 2024 14:50:23 -0400 (0:00:00.040) 0:00:01.884 ***** 13731 1727203823.65385: entering _queue_task() for managed-node3/setup 13731 1727203823.65764: worker is 1 (out of 1 available) 13731 1727203823.65879: exiting _queue_task() for managed-node3/setup 13731 1727203823.65891: done queuing things up, now waiting for results queue to drain 13731 1727203823.65892: waiting for pending results... 13731 1727203823.66100: running TaskExecutor() for managed-node3/TASK: Gather the minimum subset of ansible_facts required by the network role test 13731 1727203823.66382: in run() - task 028d2410-947f-82dc-c122-000000000026 13731 1727203823.66525: variable 'ansible_search_path' from source: unknown 13731 1727203823.66533: variable 'ansible_search_path' from source: unknown 13731 1727203823.66570: calling self._execute() 13731 1727203823.66752: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203823.66764: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203823.66780: variable 'omit' from source: magic vars 13731 1727203823.67868: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203823.72304: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203823.72486: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203823.72531: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203823.72769: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203823.72772: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203823.72986: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203823.72990: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203823.72993: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203823.73116: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203823.73137: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203823.73470: variable 'ansible_facts' from source: unknown 13731 1727203823.73597: variable 'network_test_required_facts' from source: task vars 13731 1727203823.73891: Evaluated conditional (not ansible_facts.keys() | list | intersect(network_test_required_facts) == network_test_required_facts): True 13731 1727203823.73894: variable 'omit' from source: magic vars 13731 1727203823.73897: variable 'omit' from source: magic vars 13731 1727203823.73921: variable 'omit' from source: magic vars 13731 1727203823.74063: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203823.74099: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203823.74120: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203823.74145: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203823.74165: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203823.74370: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203823.74374: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203823.74378: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203823.74515: Set connection var ansible_pipelining to False 13731 1727203823.74527: Set connection var ansible_shell_type to sh 13731 1727203823.74539: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203823.74550: Set connection var ansible_connection to ssh 13731 1727203823.74562: Set connection var ansible_shell_executable to /bin/sh 13731 1727203823.74591: Set connection var ansible_timeout to 10 13731 1727203823.74638: variable 'ansible_shell_executable' from source: unknown 13731 1727203823.74723: variable 'ansible_connection' from source: unknown 13731 1727203823.74727: variable 'ansible_module_compression' from source: unknown 13731 1727203823.74729: variable 'ansible_shell_type' from source: unknown 13731 1727203823.74732: variable 'ansible_shell_executable' from source: unknown 13731 1727203823.74734: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203823.74736: variable 'ansible_pipelining' from source: unknown 13731 1727203823.74743: variable 'ansible_timeout' from source: unknown 13731 1727203823.74752: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203823.75049: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203823.75240: variable 'omit' from source: magic vars 13731 1727203823.75244: starting attempt loop 13731 1727203823.75246: running the handler 13731 1727203823.75248: _low_level_execute_command(): starting 13731 1727203823.75250: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203823.76491: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203823.76585: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203823.76739: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203823.76757: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203823.76805: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203823.76867: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 4 <<< 13731 1727203823.79077: stdout chunk (state=3): >>>/root <<< 13731 1727203823.79282: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203823.79286: stdout chunk (state=3): >>><<< 13731 1727203823.79311: stderr chunk (state=3): >>><<< 13731 1727203823.79383: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 4 debug2: Received exit status from master 0 13731 1727203823.79397: _low_level_execute_command(): starting 13731 1727203823.79550: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701 `" && echo ansible-tmp-1727203823.7933726-13865-159361452723701="` echo /root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701 `" ) && sleep 0' 13731 1727203823.80357: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203823.80378: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203823.80394: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203823.80417: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203823.80493: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203823.80536: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203823.80562: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203823.80578: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203823.80655: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 4 <<< 13731 1727203823.83182: stdout chunk (state=3): >>>ansible-tmp-1727203823.7933726-13865-159361452723701=/root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701 <<< 13731 1727203823.83449: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203823.83453: stdout chunk (state=3): >>><<< 13731 1727203823.83456: stderr chunk (state=3): >>><<< 13731 1727203823.83459: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203823.7933726-13865-159361452723701=/root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 4 debug2: Received exit status from master 0 13731 1727203823.83578: variable 'ansible_module_compression' from source: unknown 13731 1727203823.83643: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.setup-ZIP_DEFLATED 13731 1727203823.83896: variable 'ansible_facts' from source: unknown 13731 1727203823.84137: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/AnsiballZ_setup.py 13731 1727203823.84343: Sending initial data 13731 1727203823.84359: Sent initial data (154 bytes) 13731 1727203823.85021: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203823.85094: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203823.85135: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203823.85158: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203823.85180: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203823.85245: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203823.87204: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203823.87209: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203823.87211: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/AnsiballZ_setup.py" <<< 13731 1727203823.87213: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp_tcyb8te /root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/AnsiballZ_setup.py <<< 13731 1727203823.87279: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp_tcyb8te" to remote "/root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/AnsiballZ_setup.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/AnsiballZ_setup.py" <<< 13731 1727203823.88837: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203823.88841: stdout chunk (state=3): >>><<< 13731 1727203823.88843: stderr chunk (state=3): >>><<< 13731 1727203823.88845: done transferring module to remote 13731 1727203823.88847: _low_level_execute_command(): starting 13731 1727203823.88849: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/ /root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/AnsiballZ_setup.py && sleep 0' 13731 1727203823.89663: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203823.89680: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203823.89697: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203823.89754: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203823.91502: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203823.91535: stdout chunk (state=3): >>><<< 13731 1727203823.91547: stderr chunk (state=3): >>><<< 13731 1727203823.91583: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203823.91878: _low_level_execute_command(): starting 13731 1727203823.91882: _low_level_execute_command(): executing: /bin/sh -c 'PYTHONVERBOSE=1 /usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/AnsiballZ_setup.py && sleep 0' 13731 1727203823.92724: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203823.92730: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203823.92746: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203823.92899: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203823.92936: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203823.95138: stdout chunk (state=3): >>>import _frozen_importlib # frozen import _imp # builtin import '_thread' # import '_warnings' # import '_weakref' # <<< 13731 1727203823.95143: stdout chunk (state=3): >>>import '_io' # import 'marshal' # <<< 13731 1727203823.95179: stdout chunk (state=3): >>>import 'posix' # <<< 13731 1727203823.95245: stdout chunk (state=3): >>>import '_frozen_importlib_external' # # installing zipimport hook <<< 13731 1727203823.95248: stdout chunk (state=3): >>>import 'time' # import 'zipimport' # # installed zipimport hook <<< 13731 1727203823.95363: stdout chunk (state=3): >>># /usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/encodings/__init__.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203823.95446: stdout chunk (state=3): >>>import '_codecs' # <<< 13731 1727203823.95454: stdout chunk (state=3): >>>import 'codecs' # <<< 13731 1727203823.95457: stdout chunk (state=3): >>># /usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc matches /usr/lib64/python3.12/encodings/aliases.py <<< 13731 1727203823.95462: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e5104d0> <<< 13731 1727203823.95554: stdout chunk (state=3): >>>import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e4dfb30> # /usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc' import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e512a50> import '_signal' # import '_abc' # import 'abc' # <<< 13731 1727203823.95571: stdout chunk (state=3): >>>import 'io' # import '_stat' # import 'stat' # <<< 13731 1727203823.95725: stdout chunk (state=3): >>>import '_collections_abc' # import 'genericpath' # import 'posixpath' # <<< 13731 1727203823.95812: stdout chunk (state=3): >>>import 'os' # <<< 13731 1727203823.95819: stdout chunk (state=3): >>>import '_sitebuiltins' # Processing user site-packages Processing global site-packages Adding directory: '/usr/local/lib/python3.12/site-packages' <<< 13731 1727203823.95832: stdout chunk (state=3): >>>Adding directory: '/usr/lib64/python3.12/site-packages' Adding directory: '/usr/lib/python3.12/site-packages' Processing .pth file: '/usr/lib/python3.12/site-packages/distutils-precedence.pth' # /usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8_sig.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc' import 'encodings.utf_8_sig' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2c1130> <<< 13731 1727203823.95942: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/_distutils_hack/__init__.py # code object from '/usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc' import '_distutils_hack' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2c2060> <<< 13731 1727203823.95997: stdout chunk (state=3): >>>import 'site' # Python 3.12.5 (main, Aug 23 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-1)] on linux Type "help", "copyright", "credits" or "license" for more information. <<< 13731 1727203823.96390: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc matches /usr/lib64/python3.12/base64.py # code object from '/usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc' # /usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/re/__init__.py # code object from '/usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc matches /usr/lib64/python3.12/enum.py <<< 13731 1727203823.96463: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/types.cpython-312.pyc matches /usr/lib64/python3.12/types.py <<< 13731 1727203823.96550: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/types.cpython-312.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2fff50> # /usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc matches /usr/lib64/python3.12/operator.py <<< 13731 1727203823.96667: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc' import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3140e0> # /usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc matches /usr/lib64/python3.12/functools.py <<< 13731 1727203823.96678: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc' <<< 13731 1727203823.96714: stdout chunk (state=3): >>># /usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/collections/__init__.py # code object from '/usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203823.96786: stdout chunk (state=3): >>>import 'itertools' # # /usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc matches /usr/lib64/python3.12/keyword.py # code object from '/usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e337980> <<< 13731 1727203823.96868: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc matches /usr/lib64/python3.12/reprlib.py # code object from '/usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e337f50> import '_collections' # <<< 13731 1727203823.96873: stdout chunk (state=3): >>>import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e317c20> <<< 13731 1727203823.96905: stdout chunk (state=3): >>>import '_functools' # import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e315340> <<< 13731 1727203823.96950: stdout chunk (state=3): >>>import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2fd100> <<< 13731 1727203823.97078: stdout chunk (state=3): >>># /usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc matches /usr/lib64/python3.12/re/_compiler.py <<< 13731 1727203823.97083: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc' import '_sre' # # /usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc matches /usr/lib64/python3.12/re/_parser.py <<< 13731 1727203823.97145: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc' # /usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc matches /usr/lib64/python3.12/re/_constants.py # code object from '/usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc' import 're._constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e35b950> <<< 13731 1727203823.97165: stdout chunk (state=3): >>>import 're._parser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e35a570> # /usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc matches /usr/lib64/python3.12/re/_casefix.py # code object from '/usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc' import 're._casefix' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e316210> import 're._compiler' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e358d70> <<< 13731 1727203823.97258: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc matches /usr/lib64/python3.12/copyreg.py # code object from '/usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e388950> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2fc380> <<< 13731 1727203823.97336: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc matches /usr/lib64/python3.12/struct.py # code object from '/usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e388e00> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e388cb0> <<< 13731 1727203823.97496: stdout chunk (state=3): >>># extension module 'binascii' loaded from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e3890a0> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2faea0> <<< 13731 1727203823.97527: stdout chunk (state=3): >>># /usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/importlib/__init__.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc matches /usr/lib64/python3.12/warnings.py # code object from '/usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e389760> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e389460> import 'importlib.machinery' # # /usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc matches /usr/lib64/python3.12/importlib/_abc.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc' <<< 13731 1727203823.97547: stdout chunk (state=3): >>>import 'importlib._abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e38a660> import 'importlib.util' # import 'runpy' # <<< 13731 1727203823.97569: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc matches /usr/lib64/python3.12/shutil.py <<< 13731 1727203823.97662: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc matches /usr/lib64/python3.12/fnmatch.py # code object from '/usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3a4860> import 'errno' # <<< 13731 1727203823.97733: stdout chunk (state=3): >>># extension module 'zlib' loaded from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203823.97817: stdout chunk (state=3): >>>import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e3a5fa0> # /usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc matches /usr/lib64/python3.12/bz2.py # code object from '/usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc matches /usr/lib64/python3.12/_compression.py # code object from '/usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3a6e40> <<< 13731 1727203823.97858: stdout chunk (state=3): >>># extension module '_bz2' loaded from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e3a74a0> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3a6390> # /usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc matches /usr/lib64/python3.12/lzma.py # code object from '/usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc' <<< 13731 1727203823.97886: stdout chunk (state=3): >>># extension module '_lzma' loaded from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e3a7f20> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3a7650> <<< 13731 1727203823.98015: stdout chunk (state=3): >>>import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e38a690> # /usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc matches /usr/lib64/python3.12/tempfile.py # code object from '/usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/random.cpython-312.pyc matches /usr/lib64/python3.12/random.py # code object from '/usr/lib64/python3.12/__pycache__/random.cpython-312.pyc' <<< 13731 1727203823.98071: stdout chunk (state=3): >>># extension module 'math' loaded from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0a3da0> <<< 13731 1727203823.98122: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc matches /usr/lib64/python3.12/bisect.py # code object from '/usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc' <<< 13731 1727203823.98136: stdout chunk (state=3): >>># extension module '_bisect' loaded from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0cc8f0> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0cc650> <<< 13731 1727203823.98216: stdout chunk (state=3): >>># extension module '_random' loaded from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0cc920> <<< 13731 1727203823.98241: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc matches /usr/lib64/python3.12/hashlib.py # code object from '/usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc' <<< 13731 1727203823.98262: stdout chunk (state=3): >>># extension module '_hashlib' loaded from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203823.98426: stdout chunk (state=3): >>># extension module '_hashlib' executed from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0cd250> <<< 13731 1727203823.98493: stdout chunk (state=3): >>># extension module '_blake2' loaded from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0cdc40> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0ccb00> <<< 13731 1727203823.98571: stdout chunk (state=3): >>>import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0a1f40> # /usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc matches /usr/lib64/python3.12/weakref.py # code object from '/usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc matches /usr/lib64/python3.12/_weakrefset.py <<< 13731 1727203823.98652: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0cf050> <<< 13731 1727203823.98673: stdout chunk (state=3): >>>import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0cdd90> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e38ad80> # /usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/__init__.py <<< 13731 1727203823.98732: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203823.98736: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc matches /usr/lib64/python3.12/threading.py <<< 13731 1727203823.98779: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc' <<< 13731 1727203823.98799: stdout chunk (state=3): >>>import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0f73e0> <<< 13731 1727203823.98846: stdout chunk (state=3): >>># /usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/__init__.py <<< 13731 1727203823.98870: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203823.98883: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc matches /usr/lib64/python3.12/contextlib.py <<< 13731 1727203823.98907: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc' <<< 13731 1727203823.98940: stdout chunk (state=3): >>>import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e11b7a0> <<< 13731 1727203823.98965: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc matches /usr/lib64/python3.12/pathlib.py <<< 13731 1727203823.99009: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc' <<< 13731 1727203823.99068: stdout chunk (state=3): >>>import 'ntpath' # <<< 13731 1727203823.99084: stdout chunk (state=3): >>># /usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/urllib/__init__.py # code object from '/usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc' import 'urllib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e178590> <<< 13731 1727203823.99115: stdout chunk (state=3): >>># /usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc matches /usr/lib64/python3.12/urllib/parse.py <<< 13731 1727203823.99134: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc' <<< 13731 1727203823.99159: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc matches /usr/lib64/python3.12/ipaddress.py <<< 13731 1727203823.99196: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc' <<< 13731 1727203823.99281: stdout chunk (state=3): >>>import 'ipaddress' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e17acf0> <<< 13731 1727203823.99355: stdout chunk (state=3): >>>import 'urllib.parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e1786b0> <<< 13731 1727203823.99391: stdout chunk (state=3): >>>import 'pathlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e1415b0> <<< 13731 1727203823.99425: stdout chunk (state=3): >>># /usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/glob.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc' import 'zipfile._path.glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852df81700> <<< 13731 1727203823.99451: stdout chunk (state=3): >>>import 'zipfile._path' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e11a5a0> import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0cffb0> <<< 13731 1727203823.99624: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/encodings/cp437.pyc' <<< 13731 1727203823.99635: stdout chunk (state=3): >>>import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f852e11a900> <<< 13731 1727203823.99918: stdout chunk (state=3): >>># zipimport: found 103 names in '/tmp/ansible_setup_payload__e_1mexy/ansible_setup_payload.zip' <<< 13731 1727203823.99928: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.00037: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.00069: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc matches /usr/lib64/python3.12/pkgutil.py <<< 13731 1727203824.00083: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc' <<< 13731 1727203824.00120: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc matches /usr/lib64/python3.12/typing.py <<< 13731 1727203824.00189: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc' <<< 13731 1727203824.00233: stdout chunk (state=3): >>># /usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc matches /usr/lib64/python3.12/collections/abc.py # code object from '/usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852dfe7440> <<< 13731 1727203824.00249: stdout chunk (state=3): >>>import '_typing' # <<< 13731 1727203824.00421: stdout chunk (state=3): >>>import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852dfca330> <<< 13731 1727203824.00466: stdout chunk (state=3): >>>import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852dfc9490> # zipimport: zlib available <<< 13731 1727203824.00469: stdout chunk (state=3): >>>import 'ansible' # <<< 13731 1727203824.00521: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.00524: stdout chunk (state=3): >>>import 'ansible.module_utils' # <<< 13731 1727203824.00527: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.01889: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.03297: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc matches /usr/lib64/python3.12/__future__.py # code object from '/usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852dfe5310> # /usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/json/__init__.py # code object from '/usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc matches /usr/lib64/python3.12/json/decoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc' <<< 13731 1727203824.03301: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc matches /usr/lib64/python3.12/json/scanner.py # code object from '/usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc' <<< 13731 1727203824.03337: stdout chunk (state=3): >>># extension module '_json' loaded from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e016cf0> <<< 13731 1727203824.03390: stdout chunk (state=3): >>>import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e016a80> <<< 13731 1727203824.03420: stdout chunk (state=3): >>>import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e016390> <<< 13731 1727203824.03447: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc matches /usr/lib64/python3.12/json/encoder.py <<< 13731 1727203824.03460: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc' <<< 13731 1727203824.03508: stdout chunk (state=3): >>>import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0167e0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e5129c0> <<< 13731 1727203824.03548: stdout chunk (state=3): >>>import 'atexit' # # extension module 'grp' loaded from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0179e0> <<< 13731 1727203824.03585: stdout chunk (state=3): >>># extension module 'fcntl' loaded from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e017c20> <<< 13731 1727203824.03666: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc matches /usr/lib64/python3.12/locale.py # code object from '/usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc' <<< 13731 1727203824.03677: stdout chunk (state=3): >>>import '_locale' # <<< 13731 1727203824.03736: stdout chunk (state=3): >>>import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e040110> <<< 13731 1727203824.03748: stdout chunk (state=3): >>>import 'pwd' # <<< 13731 1727203824.03768: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc matches /usr/lib64/python3.12/platform.py <<< 13731 1727203824.03801: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc' <<< 13731 1727203824.03854: stdout chunk (state=3): >>>import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d92dfa0> <<< 13731 1727203824.03901: stdout chunk (state=3): >>># extension module 'select' loaded from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.03910: stdout chunk (state=3): >>># extension module 'select' executed from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d92fbc0> # /usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc matches /usr/lib64/python3.12/selectors.py <<< 13731 1727203824.03925: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc' <<< 13731 1727203824.03980: stdout chunk (state=3): >>>import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d9304a0> # /usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc matches /usr/lib64/python3.12/shlex.py <<< 13731 1727203824.04015: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc' <<< 13731 1727203824.04038: stdout chunk (state=3): >>>import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d931370> <<< 13731 1727203824.04055: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc matches /usr/lib64/python3.12/subprocess.py <<< 13731 1727203824.04140: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc matches /usr/lib64/python3.12/signal.py <<< 13731 1727203824.04152: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc'<<< 13731 1727203824.04188: stdout chunk (state=3): >>> <<< 13731 1727203824.04251: stdout chunk (state=3): >>>import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d933fb0> <<< 13731 1727203824.04297: stdout chunk (state=3): >>># extension module '_posixsubprocess' loaded from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.04312: stdout chunk (state=3): >>> # extension module '_posixsubprocess' executed from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d9381d0> <<< 13731 1727203824.04369: stdout chunk (state=3): >>>import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d932360> # /usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc matches /usr/lib64/python3.12/traceback.py <<< 13731 1727203824.04433: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc matches /usr/lib64/python3.12/linecache.py <<< 13731 1727203824.04455: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc'<<< 13731 1727203824.04482: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc matches /usr/lib64/python3.12/tokenize.py<<< 13731 1727203824.04502: stdout chunk (state=3): >>> <<< 13731 1727203824.04634: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc'<<< 13731 1727203824.04668: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/token.cpython-312.pyc matches /usr/lib64/python3.12/token.py <<< 13731 1727203824.04692: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/token.cpython-312.pyc'<<< 13731 1727203824.04711: stdout chunk (state=3): >>> import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d93bf80><<< 13731 1727203824.04716: stdout chunk (state=3): >>> <<< 13731 1727203824.04791: stdout chunk (state=3): >>>import '_tokenize' # <<< 13731 1727203824.04837: stdout chunk (state=3): >>>import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d93aa50> <<< 13731 1727203824.04852: stdout chunk (state=3): >>>import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d93a7b0> <<< 13731 1727203824.04889: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc matches /usr/lib64/python3.12/textwrap.py <<< 13731 1727203824.04916: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc' <<< 13731 1727203824.05077: stdout chunk (state=3): >>>import 'textwrap' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d93ad20> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d932870> <<< 13731 1727203824.05121: stdout chunk (state=3): >>># extension module 'syslog' loaded from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.05140: stdout chunk (state=3): >>># extension module 'syslog' executed from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d97ff80><<< 13731 1727203824.05170: stdout chunk (state=3): >>> <<< 13731 1727203824.05194: stdout chunk (state=3): >>># /usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/__init__.py <<< 13731 1727203824.05200: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d9801a0> <<< 13731 1727203824.05235: stdout chunk (state=3): >>># /usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/journal.py <<< 13731 1727203824.05245: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc' <<< 13731 1727203824.05272: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc matches /usr/lib64/python3.12/datetime.py # code object from '/usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc' <<< 13731 1727203824.05298: stdout chunk (state=3): >>># extension module '_datetime' loaded from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.05314: stdout chunk (state=3): >>># extension module '_datetime' executed from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d981d60> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d981b20> <<< 13731 1727203824.05339: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc matches /usr/lib64/python3.12/uuid.py <<< 13731 1727203824.05365: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc' <<< 13731 1727203824.05597: stdout chunk (state=3): >>># extension module '_uuid' loaded from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d984230> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d982450> # /usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/logging/__init__.py # code object from '/usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/string.cpython-312.pyc matches /usr/lib64/python3.12/string.py <<< 13731 1727203824.05698: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/string.cpython-312.pyc' import '_string' # import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d987a10> <<< 13731 1727203824.05752: stdout chunk (state=3): >>>import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d9843e0> <<< 13731 1727203824.05809: stdout chunk (state=3): >>># extension module 'systemd._journal' loaded from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.05815: stdout chunk (state=3): >>># extension module 'systemd._journal' executed from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d988ad0> <<< 13731 1727203824.05847: stdout chunk (state=3): >>># extension module 'systemd._reader' loaded from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.05854: stdout chunk (state=3): >>># extension module 'systemd._reader' executed from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d988c50> <<< 13731 1727203824.05899: stdout chunk (state=3): >>># extension module 'systemd.id128' loaded from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.05905: stdout chunk (state=3): >>># extension module 'systemd.id128' executed from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d988b60> <<< 13731 1727203824.05924: stdout chunk (state=3): >>>import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d980470> <<< 13731 1727203824.05998: stdout chunk (state=3): >>># /usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc' <<< 13731 1727203824.06010: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc matches /usr/lib64/python3.12/socket.py <<< 13731 1727203824.06013: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc' <<< 13731 1727203824.06048: stdout chunk (state=3): >>># extension module '_socket' loaded from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.06050: stdout chunk (state=3): >>># extension module '_socket' executed from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d814350> <<< 13731 1727203824.06215: stdout chunk (state=3): >>># extension module 'array' loaded from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.06218: stdout chunk (state=3): >>>import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d815850> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d98aae0> <<< 13731 1727203824.06254: stdout chunk (state=3): >>># extension module 'systemd._daemon' loaded from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d98be90> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d98a6f0> <<< 13731 1727203824.06288: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat' # <<< 13731 1727203824.06309: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.06380: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.06491: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.06519: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.common' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.text' # <<< 13731 1727203824.06537: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.06648: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.06998: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.07608: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.08471: stdout chunk (state=3): >>>import 'ansible.module_utils.six' # import 'ansible.module_utils.six.moves' # <<< 13731 1727203824.08493: stdout chunk (state=3): >>>import 'ansible.module_utils.six.moves.collections_abc' # import 'ansible.module_utils.common.text.converters' # <<< 13731 1727203824.08541: stdout chunk (state=3): >>># /usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/__init__.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.08594: stdout chunk (state=3): >>># extension module '_ctypes' loaded from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d819a30> <<< 13731 1727203824.08709: stdout chunk (state=3): >>># /usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/_endian.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc' <<< 13731 1727203824.08743: stdout chunk (state=3): >>>import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d81a780> <<< 13731 1727203824.08748: stdout chunk (state=3): >>>import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d815b80> <<< 13731 1727203824.08804: stdout chunk (state=3): >>>import 'ansible.module_utils.compat.selinux' # <<< 13731 1727203824.08816: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.08852: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils._text' # <<< 13731 1727203824.08877: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.09092: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.09314: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc matches /usr/lib64/python3.12/copy.py # code object from '/usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc' <<< 13731 1727203824.09341: stdout chunk (state=3): >>>import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d819550> # zipimport: zlib available <<< 13731 1727203824.09834: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.10404: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.10516: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.10615: stdout chunk (state=3): >>>import 'ansible.module_utils.common.collections' # <<< 13731 1727203824.10634: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.10681: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.10726: stdout chunk (state=3): >>>import 'ansible.module_utils.common.warnings' # <<< 13731 1727203824.10739: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.10835: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.10951: stdout chunk (state=3): >>>import 'ansible.module_utils.errors' # <<< 13731 1727203824.10993: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.parsing' # <<< 13731 1727203824.11015: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.11104: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.parsing.convert_bool' # <<< 13731 1727203824.11117: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.11468: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.11825: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc matches /usr/lib64/python3.12/ast.py <<< 13731 1727203824.11906: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc' <<< 13731 1727203824.11918: stdout chunk (state=3): >>>import '_ast' # <<< 13731 1727203824.12019: stdout chunk (state=3): >>>import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d81b980> # zipimport: zlib available <<< 13731 1727203824.12141: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.12223: stdout chunk (state=3): >>>import 'ansible.module_utils.common.text.formatters' # import 'ansible.module_utils.common.validation' # import 'ansible.module_utils.common.parameters' # <<< 13731 1727203824.12262: stdout chunk (state=3): >>>import 'ansible.module_utils.common.arg_spec' # # zipimport: zlib available <<< 13731 1727203824.12322: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.12368: stdout chunk (state=3): >>>import 'ansible.module_utils.common.locale' # <<< 13731 1727203824.12389: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.12472: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.12500: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.12670: stdout chunk (state=3): >>># zipimport: zlib available # /usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/selinux/__init__.py <<< 13731 1727203824.12719: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.12828: stdout chunk (state=3): >>># extension module 'selinux._selinux' loaded from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d826240> <<< 13731 1727203824.12881: stdout chunk (state=3): >>>import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d821250> <<< 13731 1727203824.12913: stdout chunk (state=3): >>>import 'ansible.module_utils.common.file' # <<< 13731 1727203824.12925: stdout chunk (state=3): >>>import 'ansible.module_utils.common.process' # <<< 13731 1727203824.12932: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.13022: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.13143: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.13201: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/__init__.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.13231: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/distro.py <<< 13731 1727203824.13251: stdout chunk (state=3): >>># code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc' <<< 13731 1727203824.13350: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc matches /usr/lib64/python3.12/argparse.py # code object from '/usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc' <<< 13731 1727203824.13371: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc matches /usr/lib64/python3.12/gettext.py <<< 13731 1727203824.13397: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc' <<< 13731 1727203824.13477: stdout chunk (state=3): >>>import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d90eb10> <<< 13731 1727203824.13538: stdout chunk (state=3): >>>import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0467e0> <<< 13731 1727203824.13643: stdout chunk (state=3): >>>import 'distro.distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d826420> <<< 13731 1727203824.13651: stdout chunk (state=3): >>>import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d81c7a0> # destroy ansible.module_utils.distro import 'ansible.module_utils.distro' # <<< 13731 1727203824.13665: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.13704: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.13733: stdout chunk (state=3): >>>import 'ansible.module_utils.common._utils' # <<< 13731 1727203824.13740: stdout chunk (state=3): >>>import 'ansible.module_utils.common.sys_info' # <<< 13731 1727203824.13824: stdout chunk (state=3): >>>import 'ansible.module_utils.basic' # # zipimport: zlib available <<< 13731 1727203824.13859: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.modules' # # zipimport: zlib available <<< 13731 1727203824.13985: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14032: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14049: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14081: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14134: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14195: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14240: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14301: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.namespace' # # zipimport: zlib available <<< 13731 1727203824.14407: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14540: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.14578: stdout chunk (state=3): >>>import 'ansible.module_utils.compat.typing' # <<< 13731 1727203824.14596: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.14858: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.15184: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.15252: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/__init__.py <<< 13731 1727203824.15257: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.15281: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/context.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/context.py <<< 13731 1727203824.15310: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/context.cpython-312.pyc' <<< 13731 1727203824.15325: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/process.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/process.py <<< 13731 1727203824.15355: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/process.cpython-312.pyc' <<< 13731 1727203824.15378: stdout chunk (state=3): >>>import 'multiprocessing.process' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b65a0> <<< 13731 1727203824.15421: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/reduction.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/reduction.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/reduction.cpython-312.pyc' <<< 13731 1727203824.15491: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/pickle.cpython-312.pyc matches /usr/lib64/python3.12/pickle.py # code object from '/usr/lib64/python3.12/__pycache__/pickle.cpython-312.pyc' <<< 13731 1727203824.15516: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/_compat_pickle.cpython-312.pyc matches /usr/lib64/python3.12/_compat_pickle.py <<< 13731 1727203824.15543: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/_compat_pickle.cpython-312.pyc' <<< 13731 1727203824.15549: stdout chunk (state=3): >>>import '_compat_pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d40e0> <<< 13731 1727203824.15576: stdout chunk (state=3): >>># extension module '_pickle' loaded from '/usr/lib64/python3.12/lib-dynload/_pickle.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.15596: stdout chunk (state=3): >>># extension module '_pickle' executed from '/usr/lib64/python3.12/lib-dynload/_pickle.cpython-312-x86_64-linux-gnu.so' import '_pickle' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d4d46b0> <<< 13731 1727203824.15658: stdout chunk (state=3): >>>import 'pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8a32c0> <<< 13731 1727203824.15682: stdout chunk (state=3): >>>import 'multiprocessing.reduction' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b7140> <<< 13731 1727203824.15726: stdout chunk (state=3): >>>import 'multiprocessing.context' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b4c50> import 'multiprocessing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b56a0> <<< 13731 1727203824.15793: stdout chunk (state=3): >>># /usr/lib64/python3.12/multiprocessing/__pycache__/pool.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/pool.py <<< 13731 1727203824.15808: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/pool.cpython-312.pyc' <<< 13731 1727203824.15832: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/queue.cpython-312.pyc matches /usr/lib64/python3.12/queue.py <<< 13731 1727203824.15843: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/queue.cpython-312.pyc' <<< 13731 1727203824.15868: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/heapq.cpython-312.pyc matches /usr/lib64/python3.12/heapq.py <<< 13731 1727203824.15888: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/heapq.cpython-312.pyc' <<< 13731 1727203824.15907: stdout chunk (state=3): >>># extension module '_heapq' loaded from '/usr/lib64/python3.12/lib-dynload/_heapq.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.15941: stdout chunk (state=3): >>># extension module '_heapq' executed from '/usr/lib64/python3.12/lib-dynload/_heapq.cpython-312-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d4d7350> <<< 13731 1727203824.15946: stdout chunk (state=3): >>>import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d6c00> # extension module '_queue' loaded from '/usr/lib64/python3.12/lib-dynload/_queue.cpython-312-x86_64-linux-gnu.so' # extension module '_queue' executed from '/usr/lib64/python3.12/lib-dynload/_queue.cpython-312-x86_64-linux-gnu.so' import '_queue' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d4d6de0> <<< 13731 1727203824.16093: stdout chunk (state=3): >>>import 'queue' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d6030> # /usr/lib64/python3.12/multiprocessing/__pycache__/util.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/util.py <<< 13731 1727203824.16128: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/util.cpython-312.pyc' <<< 13731 1727203824.16169: stdout chunk (state=3): >>>import 'multiprocessing.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d7500> # /usr/lib64/python3.12/multiprocessing/__pycache__/connection.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/connection.py <<< 13731 1727203824.16208: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/connection.cpython-312.pyc' <<< 13731 1727203824.16277: stdout chunk (state=3): >>># extension module '_multiprocessing' loaded from '/usr/lib64/python3.12/lib-dynload/_multiprocessing.cpython-312-x86_64-linux-gnu.so' # extension module '_multiprocessing' executed from '/usr/lib64/python3.12/lib-dynload/_multiprocessing.cpython-312-x86_64-linux-gnu.so' import '_multiprocessing' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d536000> import 'multiprocessing.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d7f50> <<< 13731 1727203824.16311: stdout chunk (state=3): >>>import 'multiprocessing.pool' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b49e0> <<< 13731 1727203824.16319: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.timeout' # <<< 13731 1727203824.16332: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.collector' # <<< 13731 1727203824.16346: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.16361: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.other' # <<< 13731 1727203824.16459: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.16532: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.other.facter' # <<< 13731 1727203824.16557: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.16623: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.16692: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.other.ohai' # <<< 13731 1727203824.16715: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.16731: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.facts.system' # <<< 13731 1727203824.16747: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.16782: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.16822: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.apparmor' # <<< 13731 1727203824.16892: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.16961: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.caps' # <<< 13731 1727203824.16971: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.17027: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.17081: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.chroot' # <<< 13731 1727203824.17095: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.17172: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.17258: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.17336: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.17433: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.utils' # import 'ansible.module_utils.facts.system.cmdline' # # zipimport: zlib available <<< 13731 1727203824.18206: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.18921: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.distribution' # # zipimport: zlib available <<< 13731 1727203824.18993: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.19048: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.19079: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.19123: stdout chunk (state=3): >>>import 'ansible.module_utils.compat.datetime' # <<< 13731 1727203824.19164: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.date_time' # # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.19203: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.env' # <<< 13731 1727203824.19215: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.19293: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.19364: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.dns' # <<< 13731 1727203824.19426: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.19454: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.fips' # <<< 13731 1727203824.19472: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.19508: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.19544: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.loadavg' # <<< 13731 1727203824.19788: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/glob.py <<< 13731 1727203824.19794: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/glob.cpython-312.pyc' <<< 13731 1727203824.19820: stdout chunk (state=3): >>>import 'glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d5361e0> <<< 13731 1727203824.19849: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/configparser.cpython-312.pyc matches /usr/lib64/python3.12/configparser.py <<< 13731 1727203824.19879: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/configparser.cpython-312.pyc' <<< 13731 1727203824.20067: stdout chunk (state=3): >>>import 'configparser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d536db0> import 'ansible.module_utils.facts.system.local' # <<< 13731 1727203824.20086: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.20176: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.20267: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.lsb' # <<< 13731 1727203824.20284: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.20412: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.20541: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.pkg_mgr' # <<< 13731 1727203824.20589: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.20648: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.20755: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.platform' # <<< 13731 1727203824.20767: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.20827: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.20888: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/ssl.cpython-312.pyc matches /usr/lib64/python3.12/ssl.py <<< 13731 1727203824.21104: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/ssl.cpython-312.pyc' # extension module '_ssl' loaded from '/usr/lib64/python3.12/lib-dynload/_ssl.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.21165: stdout chunk (state=3): >>># extension module '_ssl' executed from '/usr/lib64/python3.12/lib-dynload/_ssl.cpython-312-x86_64-linux-gnu.so' import '_ssl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d5723c0><<< 13731 1727203824.21170: stdout chunk (state=3): >>> <<< 13731 1727203824.21458: stdout chunk (state=3): >>>import 'ssl' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d562180> <<< 13731 1727203824.21478: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.python' # <<< 13731 1727203824.21497: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.21588: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.21663: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.selinux' # <<< 13731 1727203824.21695: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.21815: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.21949: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.22136: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.22361: stdout chunk (state=3): >>>import 'ansible.module_utils.compat.version' # <<< 13731 1727203824.22376: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.service_mgr' # <<< 13731 1727203824.22612: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.ssh_pub_keys' # # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.22625: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/getpass.cpython-312.pyc matches /usr/lib64/python3.12/getpass.py <<< 13731 1727203824.22653: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/getpass.cpython-312.pyc' <<< 13731 1727203824.22685: stdout chunk (state=3): >>># extension module 'termios' loaded from '/usr/lib64/python3.12/lib-dynload/termios.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.22717: stdout chunk (state=3): >>># extension module 'termios' executed from '/usr/lib64/python3.12/lib-dynload/termios.cpython-312-x86_64-linux-gnu.so' import 'termios' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d585e20> <<< 13731 1727203824.22739: stdout chunk (state=3): >>>import 'getpass' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d563380> <<< 13731 1727203824.22752: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.system.user' # <<< 13731 1727203824.22779: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.22799: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.22816: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware' # <<< 13731 1727203824.22838: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.22902: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.22954: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.base' # <<< 13731 1727203824.22985: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.23228: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.23471: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.aix' # <<< 13731 1727203824.23491: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.23649: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.23804: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.23871: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.23924: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.sysctl' # <<< 13731 1727203824.23951: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.darwin' # <<< 13731 1727203824.24089: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.24239: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.24452: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.freebsd' # <<< 13731 1727203824.24472: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.dragonfly' # <<< 13731 1727203824.24495: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.24687: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.24867: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.hpux' # <<< 13731 1727203824.24895: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.24942: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.25001: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.25851: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.26655: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.linux' # <<< 13731 1727203824.26658: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.hurd' # <<< 13731 1727203824.26893: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.26989: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.netbsd' # <<< 13731 1727203824.27014: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.27161: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.27312: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.openbsd' # <<< 13731 1727203824.27331: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.27569: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.27803: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.hardware.sunos' # <<< 13731 1727203824.27829: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.27846: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.27870: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network' # <<< 13731 1727203824.27892: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.27957: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.28015: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.base' # <<< 13731 1727203824.28046: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.28190: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.28345: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.28674: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29006: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.generic_bsd' # <<< 13731 1727203824.29021: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.aix' # <<< 13731 1727203824.29044: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29097: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29293: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.darwin' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.dragonfly' # # zipimport: zlib available <<< 13731 1727203824.29333: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29437: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.fc_wwn' # <<< 13731 1727203824.29456: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29495: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29526: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.freebsd' # <<< 13731 1727203824.29558: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29639: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29728: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.hpux' # <<< 13731 1727203824.29748: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29837: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.29918: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.hurd' # <<< 13731 1727203824.29945: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.30366: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.30892: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.linux' # # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.30963: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.iscsi' # <<< 13731 1727203824.30985: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31036: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31194: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.nvme' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.netbsd' # # zipimport: zlib available <<< 13731 1727203824.31228: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31284: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.openbsd' # <<< 13731 1727203824.31300: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31425: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31542: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.network.sunos' # <<< 13731 1727203824.31570: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31589: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31692: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual' # # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.31736: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.base' # <<< 13731 1727203824.31764: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31792: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31832: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31900: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.31974: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.32085: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.32191: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.sysctl' # <<< 13731 1727203824.32289: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.freebsd' # import 'ansible.module_utils.facts.virtual.dragonfly' # # zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.32367: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.hpux' # <<< 13731 1727203824.32387: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.32701: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.33009: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.linux' # <<< 13731 1727203824.33030: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.33095: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.33165: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.netbsd' # <<< 13731 1727203824.33288: stdout chunk (state=3): >>># zipimport: zlib available # zipimport: zlib available <<< 13731 1727203824.33314: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.openbsd' # <<< 13731 1727203824.33336: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.33461: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.33577: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.virtual.sunos' # <<< 13731 1727203824.33596: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.default_collectors' # <<< 13731 1727203824.33617: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.33756: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.33879: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.ansible_collector' # <<< 13731 1727203824.33900: stdout chunk (state=3): >>>import 'ansible.module_utils.facts.compat' # <<< 13731 1727203824.33906: stdout chunk (state=3): >>>import 'ansible.module_utils.facts' # <<< 13731 1727203824.34092: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.34747: stdout chunk (state=3): >>># /usr/lib64/python3.12/encodings/__pycache__/idna.cpython-312.pyc matches /usr/lib64/python3.12/encodings/idna.py <<< 13731 1727203824.34767: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/encodings/__pycache__/idna.cpython-312.pyc' <<< 13731 1727203824.34787: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/stringprep.cpython-312.pyc matches /usr/lib64/python3.12/stringprep.py <<< 13731 1727203824.34819: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/stringprep.cpython-312.pyc' <<< 13731 1727203824.34851: stdout chunk (state=3): >>># extension module 'unicodedata' loaded from '/usr/lib64/python3.12/lib-dynload/unicodedata.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.34871: stdout chunk (state=3): >>># extension module 'unicodedata' executed from '/usr/lib64/python3.12/lib-dynload/unicodedata.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.35005: stdout chunk (state=3): >>>import 'unicodedata' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d383980> import 'stringprep' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d382420> import 'encodings.idna' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d381be0> <<< 13731 1727203824.36099: stdout chunk (state=3): >>> {"ansible_facts": {"ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQC83yKiaGAYjfqsqlfpPMYFAYI2IZVpX8WgNfvPCdI/QOKjuzs4b6SbV/Bm0ogtp9Et9STTGoHBvp3tMYQ6i0y/9DHMBxtiHYJ+rIzJ/YNNMqKc+fMqexyOsi7rKKuzfxXrYU8uPBHq5WU3WAZmJxJn+UHZGog8hUnJ8momdJG+aYo9El3Qce4gVdwORcmHZUOa49M8lLCwTovtYArmkGETUVJ+Jk8huVTzYpASWxxcw6zOvUcn52HC6dmNQv/T+k2uW6UW0rybwIrVUlZXRNODrXs8kCGgOx1OI0XYB3FndJOnORF4A9Y6onLo/zUCEaO8Pi19mcfSbo2v+bmotTVk5jcmvR3jhVYJmJE6a+dQpjSZolSqMv8mI9tkztfxM6bJlNSZcTrvZEzu7cbiE38Pp/Ku143n9iGgWNmUQ2FhUDpoWEhoA767VUunE48P8ivpVZ/u5aEyupZSLEuWEvCLmmGPVcf9hVbcXw0n8RFvUwDdD8WARVhN5GJFUVN5JM0=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHm1sMDuVWGACN5pHFDkl6tR93F90YCY4cFGcXcCoQnN+oT963FmBwTMMlfDIm4G2OUATCZuz6QFZP9trAaUzXo=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIAB71QKijTWbanEvrb0ex0kLr0wX6qyv6naldRWNiIFP", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_env": {"PYTHONVERBOSE": "1", "SHELL": "/bin/bash", "GPG_TTY": "/dev/pts/0", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.12", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.45.138 53526 10.31.47.22 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.45.138 53526 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0"}, "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_apparmor": {"statu<<< 13731 1727203824.36121: stdout chunk (state=3): >>>s": "disabled"}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "50", "second": "24", "epoch": "1727203824", "epoch_int": "1727203824", "date": "2024-09-24", "time": "14:50:24", "iso8601_micro": "2024-09-24T18:50:24.342813Z", "iso8601": "2024-09-24T18:50:24Z", "iso8601_basic": "20240924T145024342813", "iso8601_basic_short": "20240924T145024", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "10", "ansible_distribution_major_version": "10", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_python": {"version": {"major": 3, "minor": 12, "micro": 5, "releaselevel": "final", "serial": 0}, "version_info": [3, 12, 5, "final", 0], "executable": "/usr/bin/python3.12", "has_sslcontext": true, "type": "cpython"}, "ansible_system": "Linux", "ansible_kernel": "6.11.0-25.el10.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Mon Sep 16 20:35:26 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.12.5", "ansible_fqdn": "managed-node3", "ansible_hostname": "managed-node3", "ansible_nodename": "managed-node3", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "ec25272c602494034078bc876e25857f", "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_local": {}, "ansible_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_pkg_mgr": "dnf", "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_lsb": {}, "ansible_fips": false, "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "Super User", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_service_mgr": "systemd", "gather_subset": ["min"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["min"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} <<< 13731 1727203824.36955: stdout chunk (state=3): >>># clear sys.path_importer_cache <<< 13731 1727203824.37032: stdout chunk (state=3): >>># clear sys.path_hooks # clear builtins._ # clear sys.path # clear sys.argv <<< 13731 1727203824.37035: stdout chunk (state=3): >>># clear sys.ps1 # clear sys.ps2 # clear sys.last_exc # clear sys.last_type # clear sys.last_value # clear sys.last_traceback <<< 13731 1727203824.37038: stdout chunk (state=3): >>># clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys <<< 13731 1727203824.37258: stdout chunk (state=3): >>># cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # <<< 13731 1727203824.37264: stdout chunk (state=3): >>>cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro <<< 13731 1727203824.37341: stdout chunk (state=3): >>># cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules <<< 13731 1727203824.37344: stdout chunk (state=3): >>># cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing <<< 13731 1727203824.37370: stdout chunk (state=3): >>># cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context <<< 13731 1727203824.37386: stdout chunk (state=3): >>># cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _heapq <<< 13731 1727203824.37632: stdout chunk (state=3): >>># cleanup[2] removing heapq # destroy heapq # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansibl<<< 13731 1727203824.37643: stdout chunk (state=3): >>>e.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna <<< 13731 1727203824.37921: stdout chunk (state=3): >>># destroy _sitebuiltins <<< 13731 1727203824.37944: stdout chunk (state=3): >>># destroy importlib.machinery # destroy importlib._abc # destroy importlib.util <<< 13731 1727203824.37984: stdout chunk (state=3): >>># destroy _bz2 <<< 13731 1727203824.37989: stdout chunk (state=3): >>># destroy _compression # destroy _lzma <<< 13731 1727203824.38022: stdout chunk (state=3): >>># destroy _blake2 <<< 13731 1727203824.38026: stdout chunk (state=3): >>># destroy binascii # destroy zlib <<< 13731 1727203824.38043: stdout chunk (state=3): >>># destroy bz2 # destroy lzma # destroy zipfile._path <<< 13731 1727203824.38078: stdout chunk (state=3): >>># destroy zipfile # destroy pathlib # destroy zipfile._path.glob <<< 13731 1727203824.38090: stdout chunk (state=3): >>># destroy ipaddress <<< 13731 1727203824.38134: stdout chunk (state=3): >>># destroy ntpath <<< 13731 1727203824.38149: stdout chunk (state=3): >>># destroy importlib # destroy zipimport <<< 13731 1727203824.38171: stdout chunk (state=3): >>># destroy __main__ # destroy systemd.journal # destroy systemd.daemon <<< 13731 1727203824.38178: stdout chunk (state=3): >>># destroy hashlib <<< 13731 1727203824.38202: stdout chunk (state=3): >>># destroy json.decoder # destroy json.encoder <<< 13731 1727203824.38208: stdout chunk (state=3): >>># destroy json.scanner <<< 13731 1727203824.38235: stdout chunk (state=3): >>># destroy _json <<< 13731 1727203824.38246: stdout chunk (state=3): >>># destroy grp <<< 13731 1727203824.38261: stdout chunk (state=3): >>># destroy encodings <<< 13731 1727203824.38272: stdout chunk (state=3): >>># destroy _locale<<< 13731 1727203824.38280: stdout chunk (state=3): >>> <<< 13731 1727203824.38389: stdout chunk (state=3): >>># destroy locale # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog # destroy uuid # destroy selinux # destroy shutil # destroy distro # destroy distro.distro # destroy argparse # destroy logging <<< 13731 1727203824.38417: stdout chunk (state=3): >>># destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector<<< 13731 1727203824.38426: stdout chunk (state=3): >>> <<< 13731 1727203824.38445: stdout chunk (state=3): >>># destroy multiprocessing <<< 13731 1727203824.38499: stdout chunk (state=3): >>># destroy multiprocessing.connection # destroy multiprocessing.pool # destroy signal # destroy pickle # destroy multiprocessing.context # destroy array # destroy _compat_pickle # destroy _pickle # destroy queue # destroy _heapq # destroy _queue # destroy multiprocessing.process <<< 13731 1727203824.38513: stdout chunk (state=3): >>># destroy unicodedata # destroy tempfile # destroy multiprocessing.util <<< 13731 1727203824.38532: stdout chunk (state=3): >>># destroy multiprocessing.reduction <<< 13731 1727203824.38537: stdout chunk (state=3): >>># destroy selectors <<< 13731 1727203824.38702: stdout chunk (state=3): >>># destroy _multiprocessing # destroy shlex # destroy fcntl # destroy datetime # destroy subprocess # destroy base64 # destroy _ssl # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy pwd # destroy termios # destroy errno # destroy json # destroy socket # destroy struct # destroy glob # destroy fnmatch # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # cleanup[3] wiping encodings.idna<<< 13731 1727203824.38708: stdout chunk (state=3): >>> <<< 13731 1727203824.38727: stdout chunk (state=3): >>># destroy stringprep # cleanup[3] wiping configparser <<< 13731 1727203824.38735: stdout chunk (state=3): >>># cleanup[3] wiping selinux._selinux <<< 13731 1727203824.38757: stdout chunk (state=3): >>># cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc <<< 13731 1727203824.38761: stdout chunk (state=3): >>># cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser <<< 13731 1727203824.38778: stdout chunk (state=3): >>># cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket <<< 13731 1727203824.39003: stdout chunk (state=3): >>># cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform # cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser # cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types # cleanup[3] wiping encodings.utf_8_sig # cleanup[3] wiping os # destroy posixpath # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime <<< 13731 1727203824.39114: stdout chunk (state=3): >>># destroy sys.monitoring <<< 13731 1727203824.39133: stdout chunk (state=3): >>># destroy _socket <<< 13731 1727203824.39161: stdout chunk (state=3): >>># destroy _collections <<< 13731 1727203824.39202: stdout chunk (state=3): >>># destroy platform <<< 13731 1727203824.39216: stdout chunk (state=3): >>># destroy _uuid <<< 13731 1727203824.39225: stdout chunk (state=3): >>># destroy stat # destroy genericpath <<< 13731 1727203824.39240: stdout chunk (state=3): >>># destroy re._parser <<< 13731 1727203824.39247: stdout chunk (state=3): >>># destroy tokenize <<< 13731 1727203824.39275: stdout chunk (state=3): >>># destroy ansible.module_utils.six.moves.urllib <<< 13731 1727203824.39294: stdout chunk (state=3): >>># destroy copyreg # destroy contextlib <<< 13731 1727203824.39329: stdout chunk (state=3): >>># destroy _typing <<< 13731 1727203824.39351: stdout chunk (state=3): >>># destroy _tokenize <<< 13731 1727203824.39368: stdout chunk (state=3): >>># destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response <<< 13731 1727203824.39377: stdout chunk (state=3): >>># destroy ansible.module_utils.six.moves.urllib.robotparser <<< 13731 1727203824.39390: stdout chunk (state=3): >>># destroy functools # destroy operator <<< 13731 1727203824.39396: stdout chunk (state=3): >>># destroy ansible.module_utils.six.moves <<< 13731 1727203824.39421: stdout chunk (state=3): >>># destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal <<< 13731 1727203824.39456: stdout chunk (state=3): >>># clear sys.meta_path # clear sys.modules # destroy _frozen_importlib <<< 13731 1727203824.39577: stdout chunk (state=3): >>># destroy codecs # destroy encodings.aliases <<< 13731 1727203824.39597: stdout chunk (state=3): >>># destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy encodings.idna <<< 13731 1727203824.39602: stdout chunk (state=3): >>># destroy _codecs <<< 13731 1727203824.39720: stdout chunk (state=3): >>># destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections # destroy threading # destroy atexit # destroy _warnings # destroy math # destroy _bisect # destroy time # destroy _random # destroy _weakref # destroy _hashlib # destroy _operator # destroy _sre # destroy _string # destroy re # destroy itertools <<< 13731 1727203824.39769: stdout chunk (state=3): >>># destroy _abc # destroy posix <<< 13731 1727203824.39807: stdout chunk (state=3): >>># destroy _functools # destroy builtins # destroy _thread # clear sys.audit hooks <<< 13731 1727203824.40384: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203824.40387: stdout chunk (state=3): >>><<< 13731 1727203824.40390: stderr chunk (state=3): >>><<< 13731 1727203824.40636: _low_level_execute_command() done: rc=0, stdout=import _frozen_importlib # frozen import _imp # builtin import '_thread' # import '_warnings' # import '_weakref' # import '_io' # import 'marshal' # import 'posix' # import '_frozen_importlib_external' # # installing zipimport hook import 'time' # import 'zipimport' # # installed zipimport hook # /usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/encodings/__init__.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc' import '_codecs' # import 'codecs' # # /usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc matches /usr/lib64/python3.12/encodings/aliases.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e5104d0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e4dfb30> # /usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc' import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e512a50> import '_signal' # import '_abc' # import 'abc' # import 'io' # import '_stat' # import 'stat' # import '_collections_abc' # import 'genericpath' # import 'posixpath' # import 'os' # import '_sitebuiltins' # Processing user site-packages Processing global site-packages Adding directory: '/usr/local/lib/python3.12/site-packages' Adding directory: '/usr/lib64/python3.12/site-packages' Adding directory: '/usr/lib/python3.12/site-packages' Processing .pth file: '/usr/lib/python3.12/site-packages/distutils-precedence.pth' # /usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8_sig.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc' import 'encodings.utf_8_sig' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2c1130> # /usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/_distutils_hack/__init__.py # code object from '/usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc' import '_distutils_hack' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2c2060> import 'site' # Python 3.12.5 (main, Aug 23 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-1)] on linux Type "help", "copyright", "credits" or "license" for more information. # /usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc matches /usr/lib64/python3.12/base64.py # code object from '/usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc' # /usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/re/__init__.py # code object from '/usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc matches /usr/lib64/python3.12/enum.py # code object from '/usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/types.cpython-312.pyc matches /usr/lib64/python3.12/types.py # code object from '/usr/lib64/python3.12/__pycache__/types.cpython-312.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2fff50> # /usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc matches /usr/lib64/python3.12/operator.py # code object from '/usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc' import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3140e0> # /usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc matches /usr/lib64/python3.12/functools.py # code object from '/usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc' # /usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/collections/__init__.py # code object from '/usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc' import 'itertools' # # /usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc matches /usr/lib64/python3.12/keyword.py # code object from '/usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e337980> # /usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc matches /usr/lib64/python3.12/reprlib.py # code object from '/usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e337f50> import '_collections' # import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e317c20> import '_functools' # import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e315340> import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2fd100> # /usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc matches /usr/lib64/python3.12/re/_compiler.py # code object from '/usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc' import '_sre' # # /usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc matches /usr/lib64/python3.12/re/_parser.py # code object from '/usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc' # /usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc matches /usr/lib64/python3.12/re/_constants.py # code object from '/usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc' import 're._constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e35b950> import 're._parser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e35a570> # /usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc matches /usr/lib64/python3.12/re/_casefix.py # code object from '/usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc' import 're._casefix' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e316210> import 're._compiler' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e358d70> # /usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc matches /usr/lib64/python3.12/copyreg.py # code object from '/usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e388950> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2fc380> # /usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc matches /usr/lib64/python3.12/struct.py # code object from '/usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e388e00> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e388cb0> # extension module 'binascii' loaded from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e3890a0> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e2faea0> # /usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/importlib/__init__.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc matches /usr/lib64/python3.12/warnings.py # code object from '/usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e389760> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e389460> import 'importlib.machinery' # # /usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc matches /usr/lib64/python3.12/importlib/_abc.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc' import 'importlib._abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e38a660> import 'importlib.util' # import 'runpy' # # /usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc matches /usr/lib64/python3.12/shutil.py # code object from '/usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc matches /usr/lib64/python3.12/fnmatch.py # code object from '/usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3a4860> import 'errno' # # extension module 'zlib' loaded from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e3a5fa0> # /usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc matches /usr/lib64/python3.12/bz2.py # code object from '/usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc matches /usr/lib64/python3.12/_compression.py # code object from '/usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3a6e40> # extension module '_bz2' loaded from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e3a74a0> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3a6390> # /usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc matches /usr/lib64/python3.12/lzma.py # code object from '/usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc' # extension module '_lzma' loaded from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e3a7f20> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e3a7650> import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e38a690> # /usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc matches /usr/lib64/python3.12/tempfile.py # code object from '/usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/random.cpython-312.pyc matches /usr/lib64/python3.12/random.py # code object from '/usr/lib64/python3.12/__pycache__/random.cpython-312.pyc' # extension module 'math' loaded from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0a3da0> # /usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc matches /usr/lib64/python3.12/bisect.py # code object from '/usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc' # extension module '_bisect' loaded from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0cc8f0> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0cc650> # extension module '_random' loaded from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0cc920> # /usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc matches /usr/lib64/python3.12/hashlib.py # code object from '/usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc' # extension module '_hashlib' loaded from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0cd250> # extension module '_blake2' loaded from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0cdc40> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0ccb00> import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0a1f40> # /usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc matches /usr/lib64/python3.12/weakref.py # code object from '/usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc matches /usr/lib64/python3.12/_weakrefset.py # code object from '/usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0cf050> import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0cdd90> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e38ad80> # /usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/__init__.py # code object from '/usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc matches /usr/lib64/python3.12/threading.py # code object from '/usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc' import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0f73e0> # /usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/__init__.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc matches /usr/lib64/python3.12/contextlib.py # code object from '/usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc' import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e11b7a0> # /usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc matches /usr/lib64/python3.12/pathlib.py # code object from '/usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc' import 'ntpath' # # /usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/urllib/__init__.py # code object from '/usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc' import 'urllib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e178590> # /usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc matches /usr/lib64/python3.12/urllib/parse.py # code object from '/usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc matches /usr/lib64/python3.12/ipaddress.py # code object from '/usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc' import 'ipaddress' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e17acf0> import 'urllib.parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e1786b0> import 'pathlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e1415b0> # /usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/glob.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc' import 'zipfile._path.glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852df81700> import 'zipfile._path' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e11a5a0> import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0cffb0> # code object from '/usr/lib64/python3.12/encodings/cp437.pyc' import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f852e11a900> # zipimport: found 103 names in '/tmp/ansible_setup_payload__e_1mexy/ansible_setup_payload.zip' # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc matches /usr/lib64/python3.12/pkgutil.py # code object from '/usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc matches /usr/lib64/python3.12/typing.py # code object from '/usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc' # /usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc matches /usr/lib64/python3.12/collections/abc.py # code object from '/usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852dfe7440> import '_typing' # import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852dfca330> import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852dfc9490> # zipimport: zlib available import 'ansible' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc matches /usr/lib64/python3.12/__future__.py # code object from '/usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852dfe5310> # /usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/json/__init__.py # code object from '/usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc matches /usr/lib64/python3.12/json/decoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc' # /usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc matches /usr/lib64/python3.12/json/scanner.py # code object from '/usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc' # extension module '_json' loaded from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e016cf0> import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e016a80> import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e016390> # /usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc matches /usr/lib64/python3.12/json/encoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc' import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0167e0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e5129c0> import 'atexit' # # extension module 'grp' loaded from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e0179e0> # extension module 'fcntl' loaded from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852e017c20> # /usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc matches /usr/lib64/python3.12/locale.py # code object from '/usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc' import '_locale' # import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e040110> import 'pwd' # # /usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc matches /usr/lib64/python3.12/platform.py # code object from '/usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc' import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d92dfa0> # extension module 'select' loaded from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d92fbc0> # /usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc matches /usr/lib64/python3.12/selectors.py # code object from '/usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc' import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d9304a0> # /usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc matches /usr/lib64/python3.12/shlex.py # code object from '/usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc' import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d931370> # /usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc matches /usr/lib64/python3.12/subprocess.py # code object from '/usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc matches /usr/lib64/python3.12/signal.py # code object from '/usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc' import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d933fb0> # extension module '_posixsubprocess' loaded from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d9381d0> import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d932360> # /usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc matches /usr/lib64/python3.12/traceback.py # code object from '/usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc matches /usr/lib64/python3.12/linecache.py # code object from '/usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc matches /usr/lib64/python3.12/tokenize.py # code object from '/usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/token.cpython-312.pyc matches /usr/lib64/python3.12/token.py # code object from '/usr/lib64/python3.12/__pycache__/token.cpython-312.pyc' import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d93bf80> import '_tokenize' # import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d93aa50> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d93a7b0> # /usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc matches /usr/lib64/python3.12/textwrap.py # code object from '/usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc' import 'textwrap' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d93ad20> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d932870> # extension module 'syslog' loaded from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d97ff80> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d9801a0> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/journal.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc matches /usr/lib64/python3.12/datetime.py # code object from '/usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc' # extension module '_datetime' loaded from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d981d60> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d981b20> # /usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc matches /usr/lib64/python3.12/uuid.py # code object from '/usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc' # extension module '_uuid' loaded from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d984230> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d982450> # /usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/logging/__init__.py # code object from '/usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/string.cpython-312.pyc matches /usr/lib64/python3.12/string.py # code object from '/usr/lib64/python3.12/__pycache__/string.cpython-312.pyc' import '_string' # import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d987a10> import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d9843e0> # extension module 'systemd._journal' loaded from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d988ad0> # extension module 'systemd._reader' loaded from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._reader' executed from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d988c50> # extension module 'systemd.id128' loaded from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d988b60> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d980470> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc matches /usr/lib64/python3.12/socket.py # code object from '/usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc' # extension module '_socket' loaded from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d814350> # extension module 'array' loaded from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d815850> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d98aae0> # extension module 'systemd._daemon' loaded from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d98be90> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d98a6f0> # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.text' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.six' # import 'ansible.module_utils.six.moves' # import 'ansible.module_utils.six.moves.collections_abc' # import 'ansible.module_utils.common.text.converters' # # /usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/__init__.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc' # extension module '_ctypes' loaded from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d819a30> # /usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/_endian.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc' import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d81a780> import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d815b80> import 'ansible.module_utils.compat.selinux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils._text' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc matches /usr/lib64/python3.12/copy.py # code object from '/usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc' import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d819550> # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.collections' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.warnings' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.errors' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.parsing' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.parsing.convert_bool' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc matches /usr/lib64/python3.12/ast.py # code object from '/usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc' import '_ast' # import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d81b980> # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.text.formatters' # import 'ansible.module_utils.common.validation' # import 'ansible.module_utils.common.parameters' # import 'ansible.module_utils.common.arg_spec' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.locale' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/selinux/__init__.py # code object from '/usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc' # extension module 'selinux._selinux' loaded from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d826240> import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d821250> import 'ansible.module_utils.common.file' # import 'ansible.module_utils.common.process' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/__init__.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc' # /usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/distro.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc matches /usr/lib64/python3.12/argparse.py # code object from '/usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc matches /usr/lib64/python3.12/gettext.py # code object from '/usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc' import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d90eb10> import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852e0467e0> import 'distro.distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d826420> import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d81c7a0> # destroy ansible.module_utils.distro import 'ansible.module_utils.distro' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common._utils' # import 'ansible.module_utils.common.sys_info' # import 'ansible.module_utils.basic' # # zipimport: zlib available # zipimport: zlib available import 'ansible.modules' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.namespace' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat.typing' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/multiprocessing/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/__init__.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/multiprocessing/__pycache__/context.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/context.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/context.cpython-312.pyc' # /usr/lib64/python3.12/multiprocessing/__pycache__/process.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/process.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/process.cpython-312.pyc' import 'multiprocessing.process' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b65a0> # /usr/lib64/python3.12/multiprocessing/__pycache__/reduction.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/reduction.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/reduction.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/pickle.cpython-312.pyc matches /usr/lib64/python3.12/pickle.py # code object from '/usr/lib64/python3.12/__pycache__/pickle.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_compat_pickle.cpython-312.pyc matches /usr/lib64/python3.12/_compat_pickle.py # code object from '/usr/lib64/python3.12/__pycache__/_compat_pickle.cpython-312.pyc' import '_compat_pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d40e0> # extension module '_pickle' loaded from '/usr/lib64/python3.12/lib-dynload/_pickle.cpython-312-x86_64-linux-gnu.so' # extension module '_pickle' executed from '/usr/lib64/python3.12/lib-dynload/_pickle.cpython-312-x86_64-linux-gnu.so' import '_pickle' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d4d46b0> import 'pickle' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8a32c0> import 'multiprocessing.reduction' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b7140> import 'multiprocessing.context' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b4c50> import 'multiprocessing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b56a0> # /usr/lib64/python3.12/multiprocessing/__pycache__/pool.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/pool.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/pool.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/queue.cpython-312.pyc matches /usr/lib64/python3.12/queue.py # code object from '/usr/lib64/python3.12/__pycache__/queue.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/heapq.cpython-312.pyc matches /usr/lib64/python3.12/heapq.py # code object from '/usr/lib64/python3.12/__pycache__/heapq.cpython-312.pyc' # extension module '_heapq' loaded from '/usr/lib64/python3.12/lib-dynload/_heapq.cpython-312-x86_64-linux-gnu.so' # extension module '_heapq' executed from '/usr/lib64/python3.12/lib-dynload/_heapq.cpython-312-x86_64-linux-gnu.so' import '_heapq' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d4d7350> import 'heapq' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d6c00> # extension module '_queue' loaded from '/usr/lib64/python3.12/lib-dynload/_queue.cpython-312-x86_64-linux-gnu.so' # extension module '_queue' executed from '/usr/lib64/python3.12/lib-dynload/_queue.cpython-312-x86_64-linux-gnu.so' import '_queue' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d4d6de0> import 'queue' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d6030> # /usr/lib64/python3.12/multiprocessing/__pycache__/util.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/util.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/util.cpython-312.pyc' import 'multiprocessing.util' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d7500> # /usr/lib64/python3.12/multiprocessing/__pycache__/connection.cpython-312.pyc matches /usr/lib64/python3.12/multiprocessing/connection.py # code object from '/usr/lib64/python3.12/multiprocessing/__pycache__/connection.cpython-312.pyc' # extension module '_multiprocessing' loaded from '/usr/lib64/python3.12/lib-dynload/_multiprocessing.cpython-312-x86_64-linux-gnu.so' # extension module '_multiprocessing' executed from '/usr/lib64/python3.12/lib-dynload/_multiprocessing.cpython-312-x86_64-linux-gnu.so' import '_multiprocessing' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d536000> import 'multiprocessing.connection' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d4d7f50> import 'multiprocessing.pool' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d8b49e0> import 'ansible.module_utils.facts.timeout' # import 'ansible.module_utils.facts.collector' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.other' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.other.facter' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.other.ohai' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.apparmor' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.caps' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.chroot' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.utils' # import 'ansible.module_utils.facts.system.cmdline' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.distribution' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat.datetime' # import 'ansible.module_utils.facts.system.date_time' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.env' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.dns' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.fips' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.loadavg' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/glob.py # code object from '/usr/lib64/python3.12/__pycache__/glob.cpython-312.pyc' import 'glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d5361e0> # /usr/lib64/python3.12/__pycache__/configparser.cpython-312.pyc matches /usr/lib64/python3.12/configparser.py # code object from '/usr/lib64/python3.12/__pycache__/configparser.cpython-312.pyc' import 'configparser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d536db0> import 'ansible.module_utils.facts.system.local' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.lsb' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.pkg_mgr' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.platform' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/ssl.cpython-312.pyc matches /usr/lib64/python3.12/ssl.py # code object from '/usr/lib64/python3.12/__pycache__/ssl.cpython-312.pyc' # extension module '_ssl' loaded from '/usr/lib64/python3.12/lib-dynload/_ssl.cpython-312-x86_64-linux-gnu.so' # extension module '_ssl' executed from '/usr/lib64/python3.12/lib-dynload/_ssl.cpython-312-x86_64-linux-gnu.so' import '_ssl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d5723c0> import 'ssl' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d562180> import 'ansible.module_utils.facts.system.python' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.selinux' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat.version' # import 'ansible.module_utils.facts.system.service_mgr' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.system.ssh_pub_keys' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/getpass.cpython-312.pyc matches /usr/lib64/python3.12/getpass.py # code object from '/usr/lib64/python3.12/__pycache__/getpass.cpython-312.pyc' # extension module 'termios' loaded from '/usr/lib64/python3.12/lib-dynload/termios.cpython-312-x86_64-linux-gnu.so' # extension module 'termios' executed from '/usr/lib64/python3.12/lib-dynload/termios.cpython-312-x86_64-linux-gnu.so' import 'termios' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d585e20> import 'getpass' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d563380> import 'ansible.module_utils.facts.system.user' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.base' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.aix' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.sysctl' # import 'ansible.module_utils.facts.hardware.darwin' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.freebsd' # import 'ansible.module_utils.facts.hardware.dragonfly' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.hpux' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.linux' # import 'ansible.module_utils.facts.hardware.hurd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.netbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.openbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.hardware.sunos' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.base' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.generic_bsd' # import 'ansible.module_utils.facts.network.aix' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.darwin' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.dragonfly' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.fc_wwn' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.freebsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.hpux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.hurd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.linux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.iscsi' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.nvme' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.netbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.openbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.network.sunos' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.base' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.sysctl' # import 'ansible.module_utils.facts.virtual.freebsd' # import 'ansible.module_utils.facts.virtual.dragonfly' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.hpux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.linux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.netbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.openbsd' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.virtual.sunos' # import 'ansible.module_utils.facts.default_collectors' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.facts.ansible_collector' # import 'ansible.module_utils.facts.compat' # import 'ansible.module_utils.facts' # # zipimport: zlib available # /usr/lib64/python3.12/encodings/__pycache__/idna.cpython-312.pyc matches /usr/lib64/python3.12/encodings/idna.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/idna.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/stringprep.cpython-312.pyc matches /usr/lib64/python3.12/stringprep.py # code object from '/usr/lib64/python3.12/__pycache__/stringprep.cpython-312.pyc' # extension module 'unicodedata' loaded from '/usr/lib64/python3.12/lib-dynload/unicodedata.cpython-312-x86_64-linux-gnu.so' # extension module 'unicodedata' executed from '/usr/lib64/python3.12/lib-dynload/unicodedata.cpython-312-x86_64-linux-gnu.so' import 'unicodedata' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f852d383980> import 'stringprep' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d382420> import 'encodings.idna' # <_frozen_importlib_external.SourceFileLoader object at 0x7f852d381be0> {"ansible_facts": {"ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQC83yKiaGAYjfqsqlfpPMYFAYI2IZVpX8WgNfvPCdI/QOKjuzs4b6SbV/Bm0ogtp9Et9STTGoHBvp3tMYQ6i0y/9DHMBxtiHYJ+rIzJ/YNNMqKc+fMqexyOsi7rKKuzfxXrYU8uPBHq5WU3WAZmJxJn+UHZGog8hUnJ8momdJG+aYo9El3Qce4gVdwORcmHZUOa49M8lLCwTovtYArmkGETUVJ+Jk8huVTzYpASWxxcw6zOvUcn52HC6dmNQv/T+k2uW6UW0rybwIrVUlZXRNODrXs8kCGgOx1OI0XYB3FndJOnORF4A9Y6onLo/zUCEaO8Pi19mcfSbo2v+bmotTVk5jcmvR3jhVYJmJE6a+dQpjSZolSqMv8mI9tkztfxM6bJlNSZcTrvZEzu7cbiE38Pp/Ku143n9iGgWNmUQ2FhUDpoWEhoA767VUunE48P8ivpVZ/u5aEyupZSLEuWEvCLmmGPVcf9hVbcXw0n8RFvUwDdD8WARVhN5GJFUVN5JM0=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHm1sMDuVWGACN5pHFDkl6tR93F90YCY4cFGcXcCoQnN+oT963FmBwTMMlfDIm4G2OUATCZuz6QFZP9trAaUzXo=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIAB71QKijTWbanEvrb0ex0kLr0wX6qyv6naldRWNiIFP", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_env": {"PYTHONVERBOSE": "1", "SHELL": "/bin/bash", "GPG_TTY": "/dev/pts/0", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.12", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.45.138 53526 10.31.47.22 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.45.138 53526 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0"}, "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_apparmor": {"status": "disabled"}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "50", "second": "24", "epoch": "1727203824", "epoch_int": "1727203824", "date": "2024-09-24", "time": "14:50:24", "iso8601_micro": "2024-09-24T18:50:24.342813Z", "iso8601": "2024-09-24T18:50:24Z", "iso8601_basic": "20240924T145024342813", "iso8601_basic_short": "20240924T145024", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "10", "ansible_distribution_major_version": "10", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_python": {"version": {"major": 3, "minor": 12, "micro": 5, "releaselevel": "final", "serial": 0}, "version_info": [3, 12, 5, "final", 0], "executable": "/usr/bin/python3.12", "has_sslcontext": true, "type": "cpython"}, "ansible_system": "Linux", "ansible_kernel": "6.11.0-25.el10.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Mon Sep 16 20:35:26 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.12.5", "ansible_fqdn": "managed-node3", "ansible_hostname": "managed-node3", "ansible_nodename": "managed-node3", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "ec25272c602494034078bc876e25857f", "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_local": {}, "ansible_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_pkg_mgr": "dnf", "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_lsb": {}, "ansible_fips": false, "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "Super User", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_service_mgr": "systemd", "gather_subset": ["min"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["min"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} # clear sys.path_importer_cache # clear sys.path_hooks # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_exc # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _heapq # cleanup[2] removing heapq # destroy heapq # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # destroy _sitebuiltins # destroy importlib.machinery # destroy importlib._abc # destroy importlib.util # destroy _bz2 # destroy _compression # destroy _lzma # destroy _blake2 # destroy binascii # destroy zlib # destroy bz2 # destroy lzma # destroy zipfile._path # destroy zipfile # destroy pathlib # destroy zipfile._path.glob # destroy ipaddress # destroy ntpath # destroy importlib # destroy zipimport # destroy __main__ # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy grp # destroy encodings # destroy _locale # destroy locale # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog # destroy uuid # destroy selinux # destroy shutil # destroy distro # destroy distro.distro # destroy argparse # destroy logging # destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.connection # destroy multiprocessing.pool # destroy signal # destroy pickle # destroy multiprocessing.context # destroy array # destroy _compat_pickle # destroy _pickle # destroy queue # destroy _heapq # destroy _queue # destroy multiprocessing.process # destroy unicodedata # destroy tempfile # destroy multiprocessing.util # destroy multiprocessing.reduction # destroy selectors # destroy _multiprocessing # destroy shlex # destroy fcntl # destroy datetime # destroy subprocess # destroy base64 # destroy _ssl # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy pwd # destroy termios # destroy errno # destroy json # destroy socket # destroy struct # destroy glob # destroy fnmatch # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping configparser # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform # cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser # cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types # cleanup[3] wiping encodings.utf_8_sig # cleanup[3] wiping os # destroy posixpath # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy sys.monitoring # destroy _socket # destroy _collections # destroy platform # destroy _uuid # destroy stat # destroy genericpath # destroy re._parser # destroy tokenize # destroy ansible.module_utils.six.moves.urllib # destroy copyreg # destroy contextlib # destroy _typing # destroy _tokenize # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy operator # destroy ansible.module_utils.six.moves # destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal # clear sys.meta_path # clear sys.modules # destroy _frozen_importlib # destroy codecs # destroy encodings.aliases # destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy encodings.idna # destroy _codecs # destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections # destroy threading # destroy atexit # destroy _warnings # destroy math # destroy _bisect # destroy time # destroy _random # destroy _weakref # destroy _hashlib # destroy _operator # destroy _sre # destroy _string # destroy re # destroy itertools # destroy _abc # destroy posix # destroy _functools # destroy builtins # destroy _thread # clear sys.audit hooks , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. [WARNING]: Module invocation had junk after the JSON data: # clear sys.path_importer_cache # clear sys.path_hooks # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_exc # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # cleanup[2] removing ansible.module_utils.facts.namespace # cleanup[2] removing ansible.module_utils.compat.typing # cleanup[2] removing multiprocessing.process # cleanup[2] removing _compat_pickle # cleanup[2] removing _pickle # cleanup[2] removing pickle # cleanup[2] removing multiprocessing.reduction # cleanup[2] removing multiprocessing.context # cleanup[2] removing __mp_main__ # destroy __main__ # cleanup[2] removing multiprocessing # cleanup[2] removing _heapq # cleanup[2] removing heapq # destroy heapq # cleanup[2] removing _queue # cleanup[2] removing queue # cleanup[2] removing multiprocessing.util # cleanup[2] removing _multiprocessing # cleanup[2] removing multiprocessing.connection # cleanup[2] removing multiprocessing.pool # cleanup[2] removing ansible.module_utils.facts.timeout # cleanup[2] removing ansible.module_utils.facts.collector # cleanup[2] removing ansible.module_utils.facts.other # cleanup[2] removing ansible.module_utils.facts.other.facter # cleanup[2] removing ansible.module_utils.facts.other.ohai # cleanup[2] removing ansible.module_utils.facts.system # cleanup[2] removing ansible.module_utils.facts.system.apparmor # cleanup[2] removing ansible.module_utils.facts.system.caps # cleanup[2] removing ansible.module_utils.facts.system.chroot # cleanup[2] removing ansible.module_utils.facts.utils # cleanup[2] removing ansible.module_utils.facts.system.cmdline # cleanup[2] removing ansible.module_utils.facts.system.distribution # cleanup[2] removing ansible.module_utils.compat.datetime # destroy ansible.module_utils.compat.datetime # cleanup[2] removing ansible.module_utils.facts.system.date_time # cleanup[2] removing ansible.module_utils.facts.system.env # cleanup[2] removing ansible.module_utils.facts.system.dns # cleanup[2] removing ansible.module_utils.facts.system.fips # cleanup[2] removing ansible.module_utils.facts.system.loadavg # cleanup[2] removing glob # cleanup[2] removing configparser # cleanup[2] removing ansible.module_utils.facts.system.local # cleanup[2] removing ansible.module_utils.facts.system.lsb # cleanup[2] removing ansible.module_utils.facts.system.pkg_mgr # cleanup[2] removing ansible.module_utils.facts.system.platform # cleanup[2] removing _ssl # cleanup[2] removing ssl # destroy ssl # cleanup[2] removing ansible.module_utils.facts.system.python # cleanup[2] removing ansible.module_utils.facts.system.selinux # cleanup[2] removing ansible.module_utils.compat.version # destroy ansible.module_utils.compat.version # cleanup[2] removing ansible.module_utils.facts.system.service_mgr # cleanup[2] removing ansible.module_utils.facts.system.ssh_pub_keys # cleanup[2] removing termios # cleanup[2] removing getpass # cleanup[2] removing ansible.module_utils.facts.system.user # cleanup[2] removing ansible.module_utils.facts.hardware # cleanup[2] removing ansible.module_utils.facts.hardware.base # cleanup[2] removing ansible.module_utils.facts.hardware.aix # cleanup[2] removing ansible.module_utils.facts.sysctl # cleanup[2] removing ansible.module_utils.facts.hardware.darwin # cleanup[2] removing ansible.module_utils.facts.hardware.freebsd # cleanup[2] removing ansible.module_utils.facts.hardware.dragonfly # cleanup[2] removing ansible.module_utils.facts.hardware.hpux # cleanup[2] removing ansible.module_utils.facts.hardware.linux # cleanup[2] removing ansible.module_utils.facts.hardware.hurd # cleanup[2] removing ansible.module_utils.facts.hardware.netbsd # cleanup[2] removing ansible.module_utils.facts.hardware.openbsd # cleanup[2] removing ansible.module_utils.facts.hardware.sunos # cleanup[2] removing ansible.module_utils.facts.network # cleanup[2] removing ansible.module_utils.facts.network.base # cleanup[2] removing ansible.module_utils.facts.network.generic_bsd # cleanup[2] removing ansible.module_utils.facts.network.aix # cleanup[2] removing ansible.module_utils.facts.network.darwin # cleanup[2] removing ansible.module_utils.facts.network.dragonfly # cleanup[2] removing ansible.module_utils.facts.network.fc_wwn # cleanup[2] removing ansible.module_utils.facts.network.freebsd # cleanup[2] removing ansible.module_utils.facts.network.hpux # cleanup[2] removing ansible.module_utils.facts.network.hurd # cleanup[2] removing ansible.module_utils.facts.network.linux # cleanup[2] removing ansible.module_utils.facts.network.iscsi # cleanup[2] removing ansible.module_utils.facts.network.nvme # cleanup[2] removing ansible.module_utils.facts.network.netbsd # cleanup[2] removing ansible.module_utils.facts.network.openbsd # cleanup[2] removing ansible.module_utils.facts.network.sunos # cleanup[2] removing ansible.module_utils.facts.virtual # cleanup[2] removing ansible.module_utils.facts.virtual.base # cleanup[2] removing ansible.module_utils.facts.virtual.sysctl # cleanup[2] removing ansible.module_utils.facts.virtual.freebsd # cleanup[2] removing ansible.module_utils.facts.virtual.dragonfly # cleanup[2] removing ansible.module_utils.facts.virtual.hpux # cleanup[2] removing ansible.module_utils.facts.virtual.linux # cleanup[2] removing ansible.module_utils.facts.virtual.netbsd # cleanup[2] removing ansible.module_utils.facts.virtual.openbsd # cleanup[2] removing ansible.module_utils.facts.virtual.sunos # cleanup[2] removing ansible.module_utils.facts.default_collectors # cleanup[2] removing ansible.module_utils.facts.ansible_collector # cleanup[2] removing ansible.module_utils.facts.compat # cleanup[2] removing ansible.module_utils.facts # destroy ansible.module_utils.facts # destroy ansible.module_utils.facts.namespace # destroy ansible.module_utils.facts.other # destroy ansible.module_utils.facts.other.facter # destroy ansible.module_utils.facts.other.ohai # destroy ansible.module_utils.facts.system # destroy ansible.module_utils.facts.system.apparmor # destroy ansible.module_utils.facts.system.caps # destroy ansible.module_utils.facts.system.chroot # destroy ansible.module_utils.facts.system.cmdline # destroy ansible.module_utils.facts.system.distribution # destroy ansible.module_utils.facts.system.date_time # destroy ansible.module_utils.facts.system.env # destroy ansible.module_utils.facts.system.dns # destroy ansible.module_utils.facts.system.fips # destroy ansible.module_utils.facts.system.loadavg # destroy ansible.module_utils.facts.system.local # destroy ansible.module_utils.facts.system.lsb # destroy ansible.module_utils.facts.system.pkg_mgr # destroy ansible.module_utils.facts.system.platform # destroy ansible.module_utils.facts.system.python # destroy ansible.module_utils.facts.system.selinux # destroy ansible.module_utils.facts.system.service_mgr # destroy ansible.module_utils.facts.system.ssh_pub_keys # destroy ansible.module_utils.facts.system.user # destroy ansible.module_utils.facts.utils # destroy ansible.module_utils.facts.hardware # destroy ansible.module_utils.facts.hardware.base # destroy ansible.module_utils.facts.hardware.aix # destroy ansible.module_utils.facts.hardware.darwin # destroy ansible.module_utils.facts.hardware.freebsd # destroy ansible.module_utils.facts.hardware.dragonfly # destroy ansible.module_utils.facts.hardware.hpux # destroy ansible.module_utils.facts.hardware.linux # destroy ansible.module_utils.facts.hardware.hurd # destroy ansible.module_utils.facts.hardware.netbsd # destroy ansible.module_utils.facts.hardware.openbsd # destroy ansible.module_utils.facts.hardware.sunos # destroy ansible.module_utils.facts.sysctl # destroy ansible.module_utils.facts.network # destroy ansible.module_utils.facts.network.base # destroy ansible.module_utils.facts.network.generic_bsd # destroy ansible.module_utils.facts.network.aix # destroy ansible.module_utils.facts.network.darwin # destroy ansible.module_utils.facts.network.dragonfly # destroy ansible.module_utils.facts.network.fc_wwn # destroy ansible.module_utils.facts.network.freebsd # destroy ansible.module_utils.facts.network.hpux # destroy ansible.module_utils.facts.network.hurd # destroy ansible.module_utils.facts.network.linux # destroy ansible.module_utils.facts.network.iscsi # destroy ansible.module_utils.facts.network.nvme # destroy ansible.module_utils.facts.network.netbsd # destroy ansible.module_utils.facts.network.openbsd # destroy ansible.module_utils.facts.network.sunos # destroy ansible.module_utils.facts.virtual # destroy ansible.module_utils.facts.virtual.base # destroy ansible.module_utils.facts.virtual.sysctl # destroy ansible.module_utils.facts.virtual.freebsd # destroy ansible.module_utils.facts.virtual.dragonfly # destroy ansible.module_utils.facts.virtual.hpux # destroy ansible.module_utils.facts.virtual.linux # destroy ansible.module_utils.facts.virtual.netbsd # destroy ansible.module_utils.facts.virtual.openbsd # destroy ansible.module_utils.facts.virtual.sunos # destroy ansible.module_utils.facts.compat # cleanup[2] removing unicodedata # cleanup[2] removing stringprep # cleanup[2] removing encodings.idna # destroy _sitebuiltins # destroy importlib.machinery # destroy importlib._abc # destroy importlib.util # destroy _bz2 # destroy _compression # destroy _lzma # destroy _blake2 # destroy binascii # destroy zlib # destroy bz2 # destroy lzma # destroy zipfile._path # destroy zipfile # destroy pathlib # destroy zipfile._path.glob # destroy ipaddress # destroy ntpath # destroy importlib # destroy zipimport # destroy __main__ # destroy systemd.journal # destroy systemd.daemon # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy grp # destroy encodings # destroy _locale # destroy locale # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog # destroy uuid # destroy selinux # destroy shutil # destroy distro # destroy distro.distro # destroy argparse # destroy logging # destroy ansible.module_utils.facts.default_collectors # destroy ansible.module_utils.facts.ansible_collector # destroy multiprocessing # destroy multiprocessing.connection # destroy multiprocessing.pool # destroy signal # destroy pickle # destroy multiprocessing.context # destroy array # destroy _compat_pickle # destroy _pickle # destroy queue # destroy _heapq # destroy _queue # destroy multiprocessing.process # destroy unicodedata # destroy tempfile # destroy multiprocessing.util # destroy multiprocessing.reduction # destroy selectors # destroy _multiprocessing # destroy shlex # destroy fcntl # destroy datetime # destroy subprocess # destroy base64 # destroy _ssl # destroy ansible.module_utils.compat.selinux # destroy getpass # destroy pwd # destroy termios # destroy errno # destroy json # destroy socket # destroy struct # destroy glob # destroy fnmatch # destroy ansible.module_utils.compat.typing # destroy ansible.module_utils.facts.timeout # destroy ansible.module_utils.facts.collector # cleanup[3] wiping encodings.idna # destroy stringprep # cleanup[3] wiping configparser # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # destroy configparser # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform # cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser # cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types # cleanup[3] wiping encodings.utf_8_sig # cleanup[3] wiping os # destroy posixpath # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy sys.monitoring # destroy _socket # destroy _collections # destroy platform # destroy _uuid # destroy stat # destroy genericpath # destroy re._parser # destroy tokenize # destroy ansible.module_utils.six.moves.urllib # destroy copyreg # destroy contextlib # destroy _typing # destroy _tokenize # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy operator # destroy ansible.module_utils.six.moves # destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal # clear sys.meta_path # clear sys.modules # destroy _frozen_importlib # destroy codecs # destroy encodings.aliases # destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy encodings.idna # destroy _codecs # destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections # destroy threading # destroy atexit # destroy _warnings # destroy math # destroy _bisect # destroy time # destroy _random # destroy _weakref # destroy _hashlib # destroy _operator # destroy _sre # destroy _string # destroy re # destroy itertools # destroy _abc # destroy posix # destroy _functools # destroy builtins # destroy _thread # clear sys.audit hooks 13731 1727203824.42242: done with _execute_module (setup, {'gather_subset': 'min', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'setup', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203824.42245: _low_level_execute_command(): starting 13731 1727203824.42248: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203823.7933726-13865-159361452723701/ > /dev/null 2>&1 && sleep 0' 13731 1727203824.42250: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203824.42253: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203824.42255: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203824.42346: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203824.42349: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203824.42351: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203824.42353: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203824.42365: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203824.42419: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203824.42439: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203824.42463: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203824.42522: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203824.45283: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203824.45286: stdout chunk (state=3): >>><<< 13731 1727203824.45288: stderr chunk (state=3): >>><<< 13731 1727203824.45291: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203824.45293: handler run complete 13731 1727203824.45295: variable 'ansible_facts' from source: unknown 13731 1727203824.45297: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203824.45363: variable 'ansible_facts' from source: unknown 13731 1727203824.45422: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203824.45484: attempt loop complete, returning result 13731 1727203824.45492: _execute() done 13731 1727203824.45497: dumping result to json 13731 1727203824.45513: done dumping result, returning 13731 1727203824.45532: done running TaskExecutor() for managed-node3/TASK: Gather the minimum subset of ansible_facts required by the network role test [028d2410-947f-82dc-c122-000000000026] 13731 1727203824.45541: sending task result for task 028d2410-947f-82dc-c122-000000000026 13731 1727203824.45850: done sending task result for task 028d2410-947f-82dc-c122-000000000026 13731 1727203824.45854: WORKER PROCESS EXITING ok: [managed-node3] 13731 1727203824.46083: no more pending results, returning what we have 13731 1727203824.46086: results queue empty 13731 1727203824.46087: checking for any_errors_fatal 13731 1727203824.46088: done checking for any_errors_fatal 13731 1727203824.46089: checking for max_fail_percentage 13731 1727203824.46090: done checking for max_fail_percentage 13731 1727203824.46091: checking to see if all hosts have failed and the running result is not ok 13731 1727203824.46092: done checking to see if all hosts have failed 13731 1727203824.46092: getting the remaining hosts for this loop 13731 1727203824.46094: done getting the remaining hosts for this loop 13731 1727203824.46097: getting the next task for host managed-node3 13731 1727203824.46105: done getting next task for host managed-node3 13731 1727203824.46107: ^ task is: TASK: Check if system is ostree 13731 1727203824.46109: ^ state is: HOST STATE: block=2, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203824.46113: getting variables 13731 1727203824.46114: in VariableManager get_vars() 13731 1727203824.46139: Calling all_inventory to load vars for managed-node3 13731 1727203824.46142: Calling groups_inventory to load vars for managed-node3 13731 1727203824.46145: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203824.46154: Calling all_plugins_play to load vars for managed-node3 13731 1727203824.46157: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203824.46162: Calling groups_plugins_play to load vars for managed-node3 13731 1727203824.46446: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203824.46656: done with get_vars() 13731 1727203824.46669: done getting variables TASK [Check if system is ostree] *********************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:17 Tuesday 24 September 2024 14:50:24 -0400 (0:00:00.813) 0:00:02.698 ***** 13731 1727203824.46765: entering _queue_task() for managed-node3/stat 13731 1727203824.47163: worker is 1 (out of 1 available) 13731 1727203824.47180: exiting _queue_task() for managed-node3/stat 13731 1727203824.47190: done queuing things up, now waiting for results queue to drain 13731 1727203824.47192: waiting for pending results... 13731 1727203824.47347: running TaskExecutor() for managed-node3/TASK: Check if system is ostree 13731 1727203824.47487: in run() - task 028d2410-947f-82dc-c122-000000000028 13731 1727203824.47490: variable 'ansible_search_path' from source: unknown 13731 1727203824.47493: variable 'ansible_search_path' from source: unknown 13731 1727203824.47594: calling self._execute() 13731 1727203824.47625: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203824.47641: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203824.47655: variable 'omit' from source: magic vars 13731 1727203824.48191: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203824.48466: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203824.48518: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203824.48556: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203824.48685: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203824.48732: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203824.48766: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203824.48806: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203824.48842: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203824.48971: Evaluated conditional (not __network_is_ostree is defined): True 13731 1727203824.48989: variable 'omit' from source: magic vars 13731 1727203824.49035: variable 'omit' from source: magic vars 13731 1727203824.49086: variable 'omit' from source: magic vars 13731 1727203824.49121: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203824.49151: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203824.49189: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203824.49195: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203824.49207: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203824.49243: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203824.49337: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203824.49341: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203824.49374: Set connection var ansible_pipelining to False 13731 1727203824.49388: Set connection var ansible_shell_type to sh 13731 1727203824.49399: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203824.49409: Set connection var ansible_connection to ssh 13731 1727203824.49418: Set connection var ansible_shell_executable to /bin/sh 13731 1727203824.49427: Set connection var ansible_timeout to 10 13731 1727203824.49466: variable 'ansible_shell_executable' from source: unknown 13731 1727203824.49474: variable 'ansible_connection' from source: unknown 13731 1727203824.49483: variable 'ansible_module_compression' from source: unknown 13731 1727203824.49488: variable 'ansible_shell_type' from source: unknown 13731 1727203824.49493: variable 'ansible_shell_executable' from source: unknown 13731 1727203824.49554: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203824.49557: variable 'ansible_pipelining' from source: unknown 13731 1727203824.49563: variable 'ansible_timeout' from source: unknown 13731 1727203824.49565: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203824.49681: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203824.49687: variable 'omit' from source: magic vars 13731 1727203824.49773: starting attempt loop 13731 1727203824.49778: running the handler 13731 1727203824.49781: _low_level_execute_command(): starting 13731 1727203824.49783: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203824.50550: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203824.50611: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203824.50645: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203824.50658: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203824.50740: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203824.52920: stdout chunk (state=3): >>>/root <<< 13731 1727203824.53145: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203824.53148: stdout chunk (state=3): >>><<< 13731 1727203824.53150: stderr chunk (state=3): >>><<< 13731 1727203824.53181: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203824.53291: _low_level_execute_command(): starting 13731 1727203824.53298: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329 `" && echo ansible-tmp-1727203824.5319855-13899-259015273838329="` echo /root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329 `" ) && sleep 0' 13731 1727203824.53873: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203824.53947: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203824.54014: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203824.54032: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203824.54067: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203824.54185: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203824.56780: stdout chunk (state=3): >>>ansible-tmp-1727203824.5319855-13899-259015273838329=/root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329 <<< 13731 1727203824.56969: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203824.56984: stderr chunk (state=3): >>><<< 13731 1727203824.56993: stdout chunk (state=3): >>><<< 13731 1727203824.57018: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203824.5319855-13899-259015273838329=/root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203824.57078: variable 'ansible_module_compression' from source: unknown 13731 1727203824.57180: ANSIBALLZ: Using lock for stat 13731 1727203824.57182: ANSIBALLZ: Acquiring lock 13731 1727203824.57184: ANSIBALLZ: Lock acquired: 140078454805408 13731 1727203824.57186: ANSIBALLZ: Creating module 13731 1727203824.67449: ANSIBALLZ: Writing module into payload 13731 1727203824.67517: ANSIBALLZ: Writing module 13731 1727203824.67533: ANSIBALLZ: Renaming module 13731 1727203824.67536: ANSIBALLZ: Done creating module 13731 1727203824.67551: variable 'ansible_facts' from source: unknown 13731 1727203824.67608: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/AnsiballZ_stat.py 13731 1727203824.67706: Sending initial data 13731 1727203824.67710: Sent initial data (153 bytes) 13731 1727203824.68144: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203824.68149: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203824.68152: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203824.68154: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203824.68156: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203824.68209: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203824.68212: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203824.68216: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203824.68261: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203824.70496: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203824.70535: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203824.70593: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpxn9ybvzn /root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/AnsiballZ_stat.py <<< 13731 1727203824.70596: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/AnsiballZ_stat.py" <<< 13731 1727203824.70630: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpxn9ybvzn" to remote "/root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/AnsiballZ_stat.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/AnsiballZ_stat.py" <<< 13731 1727203824.71417: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203824.71567: stdout chunk (state=3): >>><<< 13731 1727203824.71570: stderr chunk (state=3): >>><<< 13731 1727203824.71572: done transferring module to remote 13731 1727203824.71584: _low_level_execute_command(): starting 13731 1727203824.71586: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/ /root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/AnsiballZ_stat.py && sleep 0' 13731 1727203824.72166: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203824.72182: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203824.72197: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203824.72265: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203824.72324: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203824.72346: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203824.72418: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203824.75028: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203824.75044: stdout chunk (state=3): >>><<< 13731 1727203824.75058: stderr chunk (state=3): >>><<< 13731 1727203824.75083: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203824.75167: _low_level_execute_command(): starting 13731 1727203824.75170: _low_level_execute_command(): executing: /bin/sh -c 'PYTHONVERBOSE=1 /usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/AnsiballZ_stat.py && sleep 0' 13731 1727203824.75754: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203824.75794: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203824.75866: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203824.78934: stdout chunk (state=3): >>>import _frozen_importlib # frozen <<< 13731 1727203824.78972: stdout chunk (state=3): >>>import _imp # builtin <<< 13731 1727203824.79028: stdout chunk (state=3): >>>import '_thread' # <<< 13731 1727203824.79040: stdout chunk (state=3): >>>import '_warnings' # import '_weakref' # <<< 13731 1727203824.79137: stdout chunk (state=3): >>>import '_io' # <<< 13731 1727203824.79194: stdout chunk (state=3): >>>import 'marshal' # <<< 13731 1727203824.79201: stdout chunk (state=3): >>>import 'posix' # <<< 13731 1727203824.79256: stdout chunk (state=3): >>>import '_frozen_importlib_external' # <<< 13731 1727203824.79265: stdout chunk (state=3): >>># installing zipimport hook <<< 13731 1727203824.79390: stdout chunk (state=3): >>>import 'time' # import 'zipimport' # # installed zipimport hook # /usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/encodings/__init__.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc' import '_codecs' # <<< 13731 1727203824.79415: stdout chunk (state=3): >>>import 'codecs' # <<< 13731 1727203824.79473: stdout chunk (state=3): >>># /usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc matches /usr/lib64/python3.12/encodings/aliases.py <<< 13731 1727203824.79499: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc' <<< 13731 1727203824.79518: stdout chunk (state=3): >>>import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c340184d0> <<< 13731 1727203824.79537: stdout chunk (state=3): >>>import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33fe7b30> <<< 13731 1727203824.79578: stdout chunk (state=3): >>># /usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8.py <<< 13731 1727203824.79607: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc' <<< 13731 1727203824.79609: stdout chunk (state=3): >>>import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3401aa50> <<< 13731 1727203824.79637: stdout chunk (state=3): >>>import '_signal' # <<< 13731 1727203824.79679: stdout chunk (state=3): >>>import '_abc' # <<< 13731 1727203824.79692: stdout chunk (state=3): >>>import 'abc' # <<< 13731 1727203824.79719: stdout chunk (state=3): >>>import 'io' # <<< 13731 1727203824.79763: stdout chunk (state=3): >>>import '_stat' # <<< 13731 1727203824.79890: stdout chunk (state=3): >>>import 'stat' # <<< 13731 1727203824.80099: stdout chunk (state=3): >>>import '_collections_abc' # <<< 13731 1727203824.80102: stdout chunk (state=3): >>> import 'genericpath' # import 'posixpath' # import 'os' # import '_sitebuiltins' # Processing user site-packages Processing global site-packages Adding directory: '/usr/local/lib/python3.12/site-packages' Adding directory: '/usr/lib64/python3.12/site-packages' Adding directory: '/usr/lib/python3.12/site-packages' Processing .pth file: '/usr/lib/python3.12/site-packages/distutils-precedence.pth' # /usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8_sig.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc' import 'encodings.utf_8_sig' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33de9130> <<< 13731 1727203824.80140: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/_distutils_hack/__init__.py <<< 13731 1727203824.80160: stdout chunk (state=3): >>># code object from '/usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.80174: stdout chunk (state=3): >>>import '_distutils_hack' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33dea060> <<< 13731 1727203824.80222: stdout chunk (state=3): >>>import 'site' # <<< 13731 1727203824.80261: stdout chunk (state=3): >>>Python 3.12.5 (main, Aug 23 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-1)] on linux Type "help", "copyright", "credits" or "license" for more information. <<< 13731 1727203824.80635: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc matches /usr/lib64/python3.12/base64.py <<< 13731 1727203824.80664: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc' <<< 13731 1727203824.80706: stdout chunk (state=3): >>># /usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/re/__init__.py <<< 13731 1727203824.80726: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.80756: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc matches /usr/lib64/python3.12/enum.py <<< 13731 1727203824.81006: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/types.cpython-312.pyc matches /usr/lib64/python3.12/types.py # code object from '/usr/lib64/python3.12/__pycache__/types.cpython-312.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e27f50> # /usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc matches /usr/lib64/python3.12/operator.py # code object from '/usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc' import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e3c0e0> # /usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc matches /usr/lib64/python3.12/functools.py # code object from '/usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc' <<< 13731 1727203824.81097: stdout chunk (state=3): >>># /usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/collections/__init__.py # code object from '/usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.81114: stdout chunk (state=3): >>>import 'itertools' # <<< 13731 1727203824.81157: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc matches /usr/lib64/python3.12/keyword.py <<< 13731 1727203824.81160: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc' <<< 13731 1727203824.81165: stdout chunk (state=3): >>>import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e5f980> <<< 13731 1727203824.81192: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc matches /usr/lib64/python3.12/reprlib.py <<< 13731 1727203824.81225: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc' <<< 13731 1727203824.81242: stdout chunk (state=3): >>>import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e5ff50> <<< 13731 1727203824.81250: stdout chunk (state=3): >>>import '_collections' # <<< 13731 1727203824.81343: stdout chunk (state=3): >>>import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e3fc20> import '_functools' # <<< 13731 1727203824.81398: stdout chunk (state=3): >>>import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e3d340> <<< 13731 1727203824.81532: stdout chunk (state=3): >>>import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e25100> <<< 13731 1727203824.81595: stdout chunk (state=3): >>># /usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc matches /usr/lib64/python3.12/re/_compiler.py <<< 13731 1727203824.81615: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc' <<< 13731 1727203824.81639: stdout chunk (state=3): >>>import '_sre' # <<< 13731 1727203824.81665: stdout chunk (state=3): >>># /usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc matches /usr/lib64/python3.12/re/_parser.py <<< 13731 1727203824.81709: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc' <<< 13731 1727203824.81740: stdout chunk (state=3): >>># /usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc matches /usr/lib64/python3.12/re/_constants.py <<< 13731 1727203824.81758: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc' <<< 13731 1727203824.81796: stdout chunk (state=3): >>>import 're._constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e83950> <<< 13731 1727203824.81897: stdout chunk (state=3): >>>import 're._parser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e82570> # /usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc matches /usr/lib64/python3.12/re/_casefix.py # code object from '/usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc' import 're._casefix' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e3e210> import 're._compiler' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e80d70> <<< 13731 1727203824.82040: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc matches /usr/lib64/python3.12/copyreg.py # code object from '/usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb0950> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e24380> # /usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc matches /usr/lib64/python3.12/struct.py # code object from '/usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33eb0e00> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb0cb0> # extension module 'binascii' loaded from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.82064: stdout chunk (state=3): >>># extension module 'binascii' executed from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33eb10a0> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e22ea0> <<< 13731 1727203824.82102: stdout chunk (state=3): >>># /usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/importlib/__init__.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.82126: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc matches /usr/lib64/python3.12/warnings.py <<< 13731 1727203824.82171: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc' <<< 13731 1727203824.82291: stdout chunk (state=3): >>>import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb1760> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb1460> import 'importlib.machinery' # # /usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc matches /usr/lib64/python3.12/importlib/_abc.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc' import 'importlib._abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb2660> import 'importlib.util' # import 'runpy' # # /usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc matches /usr/lib64/python3.12/shutil.py <<< 13731 1727203824.82328: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc' <<< 13731 1727203824.82369: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc matches /usr/lib64/python3.12/fnmatch.py <<< 13731 1727203824.82373: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc' <<< 13731 1727203824.82390: stdout chunk (state=3): >>>import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33ecc860> <<< 13731 1727203824.82404: stdout chunk (state=3): >>>import 'errno' # <<< 13731 1727203824.82431: stdout chunk (state=3): >>># extension module 'zlib' loaded from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.82452: stdout chunk (state=3): >>># extension module 'zlib' executed from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33ecdfa0> <<< 13731 1727203824.82492: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc matches /usr/lib64/python3.12/bz2.py <<< 13731 1727203824.82503: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc' <<< 13731 1727203824.82529: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc matches /usr/lib64/python3.12/_compression.py <<< 13731 1727203824.82546: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc' <<< 13731 1727203824.82571: stdout chunk (state=3): >>>import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33ecee40> <<< 13731 1727203824.82613: stdout chunk (state=3): >>># extension module '_bz2' loaded from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.82757: stdout chunk (state=3): >>># extension module '_bz2' executed from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33ecf4a0> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33ece390> # /usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc matches /usr/lib64/python3.12/lzma.py # code object from '/usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc' # extension module '_lzma' loaded from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33ecff20> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33ecf650> <<< 13731 1727203824.82791: stdout chunk (state=3): >>>import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb2690> <<< 13731 1727203824.82821: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc matches /usr/lib64/python3.12/tempfile.py <<< 13731 1727203824.82862: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc' <<< 13731 1727203824.82893: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/random.cpython-312.pyc matches /usr/lib64/python3.12/random.py <<< 13731 1727203824.82924: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/random.cpython-312.pyc' <<< 13731 1727203824.82969: stdout chunk (state=3): >>># extension module 'math' loaded from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.82977: stdout chunk (state=3): >>># extension module 'math' executed from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.82991: stdout chunk (state=3): >>>import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c5bda0> <<< 13731 1727203824.83025: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc matches /usr/lib64/python3.12/bisect.py <<< 13731 1727203824.83031: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc' <<< 13731 1727203824.83065: stdout chunk (state=3): >>># extension module '_bisect' loaded from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.83079: stdout chunk (state=3): >>># extension module '_bisect' executed from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c848f0> <<< 13731 1727203824.83100: stdout chunk (state=3): >>>import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c84650><<< 13731 1727203824.83104: stdout chunk (state=3): >>> <<< 13731 1727203824.83140: stdout chunk (state=3): >>># extension module '_random' loaded from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.83152: stdout chunk (state=3): >>># extension module '_random' executed from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c84920><<< 13731 1727203824.83195: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc matches /usr/lib64/python3.12/hashlib.py <<< 13731 1727203824.83218: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc' <<< 13731 1727203824.83307: stdout chunk (state=3): >>># extension module '_hashlib' loaded from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.83495: stdout chunk (state=3): >>> <<< 13731 1727203824.83512: stdout chunk (state=3): >>># extension module '_hashlib' executed from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.83520: stdout chunk (state=3): >>>import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c85250> <<< 13731 1727203824.83694: stdout chunk (state=3): >>># extension module '_blake2' loaded from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.83724: stdout chunk (state=3): >>># extension module '_blake2' executed from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.83727: stdout chunk (state=3): >>>import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c85c40> <<< 13731 1727203824.83757: stdout chunk (state=3): >>>import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c84b00> <<< 13731 1727203824.83791: stdout chunk (state=3): >>>import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c59f40> <<< 13731 1727203824.83827: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc matches /usr/lib64/python3.12/weakref.py <<< 13731 1727203824.83866: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc' <<< 13731 1727203824.83900: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc matches /usr/lib64/python3.12/_weakrefset.py <<< 13731 1727203824.83926: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc' <<< 13731 1727203824.83963: stdout chunk (state=3): >>>import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c87050> <<< 13731 1727203824.84015: stdout chunk (state=3): >>>import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c85d90> <<< 13731 1727203824.84055: stdout chunk (state=3): >>>import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb2d80> <<< 13731 1727203824.84081: stdout chunk (state=3): >>># /usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/__init__.py <<< 13731 1727203824.84167: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.84195: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc matches /usr/lib64/python3.12/threading.py<<< 13731 1727203824.84241: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc' <<< 13731 1727203824.84288: stdout chunk (state=3): >>>import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33caf3e0> <<< 13731 1727203824.84368: stdout chunk (state=3): >>># /usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/__init__.py <<< 13731 1727203824.84388: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc'<<< 13731 1727203824.84420: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc matches /usr/lib64/python3.12/contextlib.py <<< 13731 1727203824.84461: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc' <<< 13731 1727203824.84550: stdout chunk (state=3): >>>import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33cd37a0> # /usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc matches /usr/lib64/python3.12/pathlib.py <<< 13731 1727203824.84616: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc' <<< 13731 1727203824.84748: stdout chunk (state=3): >>>import 'ntpath' # # /usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/urllib/__init__.py<<< 13731 1727203824.84751: stdout chunk (state=3): >>> <<< 13731 1727203824.84763: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.84786: stdout chunk (state=3): >>>import 'urllib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33d30590> # /usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc matches /usr/lib64/python3.12/urllib/parse.py <<< 13731 1727203824.84834: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc'<<< 13731 1727203824.84874: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc matches /usr/lib64/python3.12/ipaddress.py <<< 13731 1727203824.84940: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc' <<< 13731 1727203824.85079: stdout chunk (state=3): >>>import 'ipaddress' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33d32cf0><<< 13731 1727203824.85192: stdout chunk (state=3): >>> import 'urllib.parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33d306b0><<< 13731 1727203824.85198: stdout chunk (state=3): >>> <<< 13731 1727203824.85256: stdout chunk (state=3): >>>import 'pathlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33cf95b0><<< 13731 1727203824.85261: stdout chunk (state=3): >>> <<< 13731 1727203824.85296: stdout chunk (state=3): >>># /usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/glob.py<<< 13731 1727203824.85311: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc'<<< 13731 1727203824.85329: stdout chunk (state=3): >>> import 'zipfile._path.glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b39700><<< 13731 1727203824.85331: stdout chunk (state=3): >>> <<< 13731 1727203824.85367: stdout chunk (state=3): >>>import 'zipfile._path' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33cd25a0><<< 13731 1727203824.85377: stdout chunk (state=3): >>> import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c87fb0><<< 13731 1727203824.85387: stdout chunk (state=3): >>> <<< 13731 1727203824.85549: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/encodings/cp437.pyc'<<< 13731 1727203824.85565: stdout chunk (state=3): >>> <<< 13731 1727203824.85596: stdout chunk (state=3): >>>import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f0c33cd2900><<< 13731 1727203824.85599: stdout chunk (state=3): >>> <<< 13731 1727203824.85789: stdout chunk (state=3): >>># zipimport: found 30 names in '/tmp/ansible_stat_payload_alr9sf9p/ansible_stat_payload.zip'<<< 13731 1727203824.85810: stdout chunk (state=3): >>> # zipimport: zlib available<<< 13731 1727203824.85989: stdout chunk (state=3): >>> <<< 13731 1727203824.86032: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.86074: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc matches /usr/lib64/python3.12/pkgutil.py <<< 13731 1727203824.86102: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc' <<< 13731 1727203824.86162: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc matches /usr/lib64/python3.12/typing.py <<< 13731 1727203824.86270: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc' <<< 13731 1727203824.86326: stdout chunk (state=3): >>># /usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc matches /usr/lib64/python3.12/collections/abc.py # code object from '/usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc'<<< 13731 1727203824.86330: stdout chunk (state=3): >>> <<< 13731 1727203824.86351: stdout chunk (state=3): >>>import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b8b470> <<< 13731 1727203824.86365: stdout chunk (state=3): >>>import '_typing' # <<< 13731 1727203824.86619: stdout chunk (state=3): >>>import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b6e360><<< 13731 1727203824.86638: stdout chunk (state=3): >>> <<< 13731 1727203824.86651: stdout chunk (state=3): >>>import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b6d4c0> <<< 13731 1727203824.86668: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.86673: stdout chunk (state=3): >>> <<< 13731 1727203824.86708: stdout chunk (state=3): >>>import 'ansible' # <<< 13731 1727203824.86712: stdout chunk (state=3): >>> <<< 13731 1727203824.86746: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.86779: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.86783: stdout chunk (state=3): >>> <<< 13731 1727203824.86806: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.86830: stdout chunk (state=3): >>> import 'ansible.module_utils' # <<< 13731 1727203824.86864: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.86986: stdout chunk (state=3): >>> <<< 13731 1727203824.89034: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.90778: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc matches /usr/lib64/python3.12/__future__.py <<< 13731 1727203824.90795: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc' <<< 13731 1727203824.90820: stdout chunk (state=3): >>>import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b89340> <<< 13731 1727203824.90857: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/json/__init__.py <<< 13731 1727203824.90879: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc'<<< 13731 1727203824.90885: stdout chunk (state=3): >>> <<< 13731 1727203824.90919: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc matches /usr/lib64/python3.12/json/decoder.py <<< 13731 1727203824.90944: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc' <<< 13731 1727203824.90979: stdout chunk (state=3): >>># /usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc matches /usr/lib64/python3.12/json/scanner.py<<< 13731 1727203824.90993: stdout chunk (state=3): >>> <<< 13731 1727203824.91005: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc' <<< 13731 1727203824.91054: stdout chunk (state=3): >>># extension module '_json' loaded from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.91114: stdout chunk (state=3): >>> import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33bb6de0> import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bb6b70><<< 13731 1727203824.91119: stdout chunk (state=3): >>> <<< 13731 1727203824.91302: stdout chunk (state=3): >>>import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bb6480> # /usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc matches /usr/lib64/python3.12/json/encoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc' import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bb6ed0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3401a9c0> import 'atexit' # # extension module 'grp' loaded from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33bb7b00> # extension module 'fcntl' loaded from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.91317: stdout chunk (state=3): >>># extension module 'fcntl' executed from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.91320: stdout chunk (state=3): >>>import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33bb7d40> <<< 13731 1727203824.91417: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc matches /usr/lib64/python3.12/locale.py # code object from '/usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc' <<< 13731 1727203824.91449: stdout chunk (state=3): >>>import '_locale' # <<< 13731 1727203824.91518: stdout chunk (state=3): >>>import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bdc230> <<< 13731 1727203824.91545: stdout chunk (state=3): >>>import 'pwd' # <<< 13731 1727203824.91583: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc matches /usr/lib64/python3.12/platform.py <<< 13731 1727203824.91622: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc'<<< 13731 1727203824.91678: stdout chunk (state=3): >>> import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33511f10><<< 13731 1727203824.91685: stdout chunk (state=3): >>> <<< 13731 1727203824.91719: stdout chunk (state=3): >>># extension module 'select' loaded from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.91740: stdout chunk (state=3): >>> <<< 13731 1727203824.91746: stdout chunk (state=3): >>># extension module 'select' executed from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.91756: stdout chunk (state=3): >>>import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33513b30><<< 13731 1727203824.91765: stdout chunk (state=3): >>> <<< 13731 1727203824.91815: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc matches /usr/lib64/python3.12/selectors.py # code object from '/usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc'<<< 13731 1727203824.91818: stdout chunk (state=3): >>> <<< 13731 1727203824.91867: stdout chunk (state=3): >>>import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33514500><<< 13731 1727203824.91902: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc matches /usr/lib64/python3.12/shlex.py <<< 13731 1727203824.91947: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc'<<< 13731 1727203824.91984: stdout chunk (state=3): >>> import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335153d0><<< 13731 1727203824.91992: stdout chunk (state=3): >>> <<< 13731 1727203824.92023: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc matches /usr/lib64/python3.12/subprocess.py<<< 13731 1727203824.92078: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc' <<< 13731 1727203824.92120: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc matches /usr/lib64/python3.12/signal.py <<< 13731 1727203824.92135: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc' <<< 13731 1727203824.92214: stdout chunk (state=3): >>>import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335200e0><<< 13731 1727203824.92273: stdout chunk (state=3): >>> # extension module '_posixsubprocess' loaded from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.92284: stdout chunk (state=3): >>># extension module '_posixsubprocess' executed from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.92297: stdout chunk (state=3): >>>import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33520230> <<< 13731 1727203824.92323: stdout chunk (state=3): >>>import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335163c0><<< 13731 1727203824.92354: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc matches /usr/lib64/python3.12/traceback.py <<< 13731 1727203824.92395: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc'<<< 13731 1727203824.92435: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc matches /usr/lib64/python3.12/linecache.py<<< 13731 1727203824.92439: stdout chunk (state=3): >>> <<< 13731 1727203824.92452: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc' <<< 13731 1727203824.92476: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc matches /usr/lib64/python3.12/tokenize.py<<< 13731 1727203824.92516: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc' <<< 13731 1727203824.92555: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/token.cpython-312.pyc matches /usr/lib64/python3.12/token.py <<< 13731 1727203824.92575: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/token.cpython-312.pyc' <<< 13731 1727203824.92604: stdout chunk (state=3): >>>import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33523f80> import '_tokenize' # <<< 13731 1727203824.92609: stdout chunk (state=3): >>> <<< 13731 1727203824.92696: stdout chunk (state=3): >>>import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33522a50> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335227b0><<< 13731 1727203824.92725: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc matches /usr/lib64/python3.12/textwrap.py <<< 13731 1727203824.92746: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc'<<< 13731 1727203824.92857: stdout chunk (state=3): >>> import 'textwrap' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33522d20><<< 13731 1727203824.92863: stdout chunk (state=3): >>> <<< 13731 1727203824.92909: stdout chunk (state=3): >>>import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335168d0><<< 13731 1727203824.92914: stdout chunk (state=3): >>> <<< 13731 1727203824.92953: stdout chunk (state=3): >>># extension module 'syslog' loaded from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.92958: stdout chunk (state=3): >>> <<< 13731 1727203824.92984: stdout chunk (state=3): >>># extension module 'syslog' executed from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.92987: stdout chunk (state=3): >>>import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c335681a0><<< 13731 1727203824.92990: stdout chunk (state=3): >>> <<< 13731 1727203824.93027: stdout chunk (state=3): >>># /usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/__init__.py<<< 13731 1727203824.93039: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc'<<< 13731 1727203824.93051: stdout chunk (state=3): >>> <<< 13731 1727203824.93064: stdout chunk (state=3): >>>import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33568350> <<< 13731 1727203824.93098: stdout chunk (state=3): >>># /usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/journal.py<<< 13731 1727203824.93129: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc' <<< 13731 1727203824.93164: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc matches /usr/lib64/python3.12/datetime.py<<< 13731 1727203824.93167: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc'<<< 13731 1727203824.93233: stdout chunk (state=3): >>> # extension module '_datetime' loaded from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.93245: stdout chunk (state=3): >>># extension module '_datetime' executed from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.93253: stdout chunk (state=3): >>>import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33569df0><<< 13731 1727203824.93259: stdout chunk (state=3): >>> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33569bb0><<< 13731 1727203824.93294: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc matches /usr/lib64/python3.12/uuid.py <<< 13731 1727203824.93465: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc'<<< 13731 1727203824.93532: stdout chunk (state=3): >>> # extension module '_uuid' loaded from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.93539: stdout chunk (state=3): >>> # extension module '_uuid' executed from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.93563: stdout chunk (state=3): >>> import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c3356c2f0> <<< 13731 1727203824.93581: stdout chunk (state=3): >>>import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3356a4e0> <<< 13731 1727203824.93611: stdout chunk (state=3): >>># /usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/logging/__init__.py<<< 13731 1727203824.93674: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.93706: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/string.cpython-312.pyc matches /usr/lib64/python3.12/string.py <<< 13731 1727203824.93733: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/string.cpython-312.pyc' <<< 13731 1727203824.93762: stdout chunk (state=3): >>>import '_string' # <<< 13731 1727203824.93832: stdout chunk (state=3): >>>import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3356faa0> <<< 13731 1727203824.94118: stdout chunk (state=3): >>>import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3356c4a0> # extension module 'systemd._journal' loaded from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.94122: stdout chunk (state=3): >>># extension module 'systemd._journal' executed from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.94142: stdout chunk (state=3): >>>import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c335705c0> <<< 13731 1727203824.94199: stdout chunk (state=3): >>># extension module 'systemd._reader' loaded from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.94220: stdout chunk (state=3): >>># extension module 'systemd._reader' executed from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33570e90><<< 13731 1727203824.94227: stdout chunk (state=3): >>> <<< 13731 1727203824.94293: stdout chunk (state=3): >>># extension module 'systemd.id128' loaded from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33570920><<< 13731 1727203824.94321: stdout chunk (state=3): >>> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33568470> <<< 13731 1727203824.94362: stdout chunk (state=3): >>># /usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc'<<< 13731 1727203824.94400: stdout chunk (state=3): >>> # /usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc matches /usr/lib64/python3.12/socket.py<<< 13731 1727203824.94407: stdout chunk (state=3): >>> <<< 13731 1727203824.94437: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc'<<< 13731 1727203824.94479: stdout chunk (state=3): >>> # extension module '_socket' loaded from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.94484: stdout chunk (state=3): >>> <<< 13731 1727203824.94534: stdout chunk (state=3): >>># extension module '_socket' executed from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.94540: stdout chunk (state=3): >>>import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c335fc260><<< 13731 1727203824.94688: stdout chunk (state=3): >>> <<< 13731 1727203824.94793: stdout chunk (state=3): >>># extension module 'array' loaded from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.94825: stdout chunk (state=3): >>># extension module 'array' executed from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.94828: stdout chunk (state=3): >>>import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c335fd520> <<< 13731 1727203824.94860: stdout chunk (state=3): >>>import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33572a20> <<< 13731 1727203824.94897: stdout chunk (state=3): >>># extension module 'systemd._daemon' loaded from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.94915: stdout chunk (state=3): >>> # extension module 'systemd._daemon' executed from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203824.94918: stdout chunk (state=3): >>> import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33573da0> <<< 13731 1727203824.94939: stdout chunk (state=3): >>>import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33572630><<< 13731 1727203824.94944: stdout chunk (state=3): >>> <<< 13731 1727203824.94970: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.94996: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.95010: stdout chunk (state=3): >>> <<< 13731 1727203824.95015: stdout chunk (state=3): >>>import 'ansible.module_utils.compat' # <<< 13731 1727203824.95042: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.95047: stdout chunk (state=3): >>> <<< 13731 1727203824.95181: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.95184: stdout chunk (state=3): >>> <<< 13731 1727203824.95305: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.95327: stdout chunk (state=3): >>> # zipimport: zlib available <<< 13731 1727203824.95357: stdout chunk (state=3): >>>import 'ansible.module_utils.common' # # zipimport: zlib available <<< 13731 1727203824.95384: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.95395: stdout chunk (state=3): >>> import 'ansible.module_utils.common.text' # <<< 13731 1727203824.95420: stdout chunk (state=3): >>> # zipimport: zlib available <<< 13731 1727203824.95603: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.95800: stdout chunk (state=3): >>> # zipimport: zlib available <<< 13731 1727203824.96695: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.97566: stdout chunk (state=3): >>>import 'ansible.module_utils.six' # <<< 13731 1727203824.97591: stdout chunk (state=3): >>>import 'ansible.module_utils.six.moves' # <<< 13731 1727203824.97614: stdout chunk (state=3): >>>import 'ansible.module_utils.six.moves.collections_abc' # <<< 13731 1727203824.97636: stdout chunk (state=3): >>>import 'ansible.module_utils.common.text.converters' # <<< 13731 1727203824.97679: stdout chunk (state=3): >>># /usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/__init__.py <<< 13731 1727203824.97714: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203824.97784: stdout chunk (state=3): >>># extension module '_ctypes' loaded from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.97814: stdout chunk (state=3): >>># extension module '_ctypes' executed from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' <<< 13731 1727203824.97818: stdout chunk (state=3): >>>import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33401820><<< 13731 1727203824.97823: stdout chunk (state=3): >>> <<< 13731 1727203824.97937: stdout chunk (state=3): >>># /usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/_endian.py<<< 13731 1727203824.97951: stdout chunk (state=3): >>> <<< 13731 1727203824.97956: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc' <<< 13731 1727203824.97995: stdout chunk (state=3): >>>import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c334027b0> <<< 13731 1727203824.98013: stdout chunk (state=3): >>>import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335fd760><<< 13731 1727203824.98080: stdout chunk (state=3): >>> import 'ansible.module_utils.compat.selinux' # <<< 13731 1727203824.98111: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.98113: stdout chunk (state=3): >>> <<< 13731 1727203824.98141: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.98146: stdout chunk (state=3): >>> <<< 13731 1727203824.98168: stdout chunk (state=3): >>>import 'ansible.module_utils._text' # <<< 13731 1727203824.98173: stdout chunk (state=3): >>> <<< 13731 1727203824.98196: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.98292: stdout chunk (state=3): >>> <<< 13731 1727203824.98440: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203824.98445: stdout chunk (state=3): >>> <<< 13731 1727203824.98678: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc matches /usr/lib64/python3.12/copy.py<<< 13731 1727203824.98697: stdout chunk (state=3): >>> <<< 13731 1727203824.98700: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc'<<< 13731 1727203824.98703: stdout chunk (state=3): >>> <<< 13731 1727203824.98730: stdout chunk (state=3): >>>import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33402c60> <<< 13731 1727203824.98759: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203824.99484: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.00212: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.00324: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203825.00439: stdout chunk (state=3): >>> import 'ansible.module_utils.common.collections' # <<< 13731 1727203825.00470: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.00534: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.00616: stdout chunk (state=3): >>>import 'ansible.module_utils.common.warnings' # <<< 13731 1727203825.00630: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.00740: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203825.00752: stdout chunk (state=3): >>> <<< 13731 1727203825.00877: stdout chunk (state=3): >>>import 'ansible.module_utils.errors' # <<< 13731 1727203825.00917: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.00971: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.parsing' # <<< 13731 1727203825.00974: stdout chunk (state=3): >>> # zipimport: zlib available <<< 13731 1727203825.01097: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.module_utils.parsing.convert_bool' # <<< 13731 1727203825.01100: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.01527: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.01851: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc matches /usr/lib64/python3.12/ast.py<<< 13731 1727203825.01900: stdout chunk (state=3): >>> <<< 13731 1727203825.01948: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc'<<< 13731 1727203825.02110: stdout chunk (state=3): >>> import '_ast' # import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c334039e0> # zipimport: zlib available<<< 13731 1727203825.02225: stdout chunk (state=3): >>> # zipimport: zlib available <<< 13731 1727203825.02335: stdout chunk (state=3): >>>import 'ansible.module_utils.common.text.formatters' # <<< 13731 1727203825.02362: stdout chunk (state=3): >>> import 'ansible.module_utils.common.validation' # import 'ansible.module_utils.common.parameters' # <<< 13731 1727203825.02388: stdout chunk (state=3): >>>import 'ansible.module_utils.common.arg_spec' # <<< 13731 1727203825.02453: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.02482: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203825.02579: stdout chunk (state=3): >>> <<< 13731 1727203825.02605: stdout chunk (state=3): >>>import 'ansible.module_utils.common.locale' # # zipimport: zlib available <<< 13731 1727203825.02686: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.02711: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.02889: stdout chunk (state=3): >>># zipimport: zlib available # /usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/selinux/__init__.py <<< 13731 1727203825.02950: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203825.03056: stdout chunk (state=3): >>># extension module 'selinux._selinux' loaded from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203825.03083: stdout chunk (state=3): >>> # extension module 'selinux._selinux' executed from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so'<<< 13731 1727203825.03088: stdout chunk (state=3): >>> import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c3340e3c0> <<< 13731 1727203825.03138: stdout chunk (state=3): >>>import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3340b650><<< 13731 1727203825.03143: stdout chunk (state=3): >>> <<< 13731 1727203825.03191: stdout chunk (state=3): >>>import 'ansible.module_utils.common.file' # <<< 13731 1727203825.03210: stdout chunk (state=3): >>>import 'ansible.module_utils.common.process' # <<< 13731 1727203825.03218: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203825.03317: stdout chunk (state=3): >>> # zipimport: zlib available <<< 13731 1727203825.03431: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.03479: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.03552: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/__init__.py <<< 13731 1727203825.03566: stdout chunk (state=3): >>># code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc' <<< 13731 1727203825.03605: stdout chunk (state=3): >>># /usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/distro.py <<< 13731 1727203825.03647: stdout chunk (state=3): >>># code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc' <<< 13731 1727203825.03680: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc matches /usr/lib64/python3.12/argparse.py<<< 13731 1727203825.03767: stdout chunk (state=3): >>> # code object from '/usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc'<<< 13731 1727203825.03773: stdout chunk (state=3): >>> <<< 13731 1727203825.03802: stdout chunk (state=3): >>># /usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc matches /usr/lib64/python3.12/gettext.py<<< 13731 1727203825.03807: stdout chunk (state=3): >>> <<< 13731 1727203825.03838: stdout chunk (state=3): >>># code object from '/usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc'<<< 13731 1727203825.03844: stdout chunk (state=3): >>> <<< 13731 1727203825.03931: stdout chunk (state=3): >>>import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33be2ba0><<< 13731 1727203825.03936: stdout chunk (state=3): >>> <<< 13731 1727203825.04001: stdout chunk (state=3): >>>import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bfe870><<< 13731 1727203825.04006: stdout chunk (state=3): >>> <<< 13731 1727203825.04117: stdout chunk (state=3): >>>import 'distro.distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3340e480><<< 13731 1727203825.04135: stdout chunk (state=3): >>> <<< 13731 1727203825.04143: stdout chunk (state=3): >>>import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c334031d0> <<< 13731 1727203825.04149: stdout chunk (state=3): >>># destroy ansible.module_utils.distro<<< 13731 1727203825.04184: stdout chunk (state=3): >>> import 'ansible.module_utils.distro' # # zipimport: zlib available <<< 13731 1727203825.04227: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203825.04270: stdout chunk (state=3): >>> import 'ansible.module_utils.common._utils' # <<< 13731 1727203825.04287: stdout chunk (state=3): >>> <<< 13731 1727203825.04290: stdout chunk (state=3): >>>import 'ansible.module_utils.common.sys_info' # <<< 13731 1727203825.04369: stdout chunk (state=3): >>>import 'ansible.module_utils.basic' # <<< 13731 1727203825.04397: stdout chunk (state=3): >>> # zipimport: zlib available <<< 13731 1727203825.04427: stdout chunk (state=3): >>># zipimport: zlib available import 'ansible.modules' # <<< 13731 1727203825.04454: stdout chunk (state=3): >>> # zipimport: zlib available<<< 13731 1727203825.04460: stdout chunk (state=3): >>> <<< 13731 1727203825.04671: stdout chunk (state=3): >>># zipimport: zlib available<<< 13731 1727203825.04881: stdout chunk (state=3): >>> <<< 13731 1727203825.05107: stdout chunk (state=3): >>># zipimport: zlib available <<< 13731 1727203825.05123: stdout chunk (state=3): >>> <<< 13731 1727203825.05143: stdout chunk (state=3): >>>{"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/run/ostree-booted", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 13731 1727203825.05174: stdout chunk (state=3): >>># destroy __main__ <<< 13731 1727203825.05650: stdout chunk (state=3): >>># clear sys.path_importer_cache <<< 13731 1727203825.05683: stdout chunk (state=3): >>># clear sys.path_hooks # clear builtins._<<< 13731 1727203825.05690: stdout chunk (state=3): >>> # clear sys.path<<< 13731 1727203825.05739: stdout chunk (state=3): >>> # clear sys.argv # clear sys.ps1 # clear sys.ps2 <<< 13731 1727203825.05757: stdout chunk (state=3): >>># clear sys.last_exc # clear sys.last_type # clear sys.last_value<<< 13731 1727203825.05773: stdout chunk (state=3): >>> # clear sys.last_traceback # clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys<<< 13731 1727203825.05815: stdout chunk (state=3): >>> # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs<<< 13731 1727203825.05822: stdout chunk (state=3): >>> # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath<<< 13731 1727203825.05871: stdout chunk (state=3): >>> # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword<<< 13731 1727203825.05881: stdout chunk (state=3): >>> # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct<<< 13731 1727203825.05894: stdout chunk (state=3): >>> # cleanup[2] removing binascii # cleanup[2] removing base64 # destroy base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc<<< 13731 1727203825.06021: stdout chunk (state=3): >>> # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # destroy socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules <<< 13731 1727203825.06372: stdout chunk (state=3): >>># destroy _sitebuiltins # destroy importlib.machinery<<< 13731 1727203825.06414: stdout chunk (state=3): >>> # destroy importlib._abc # destroy importlib.util<<< 13731 1727203825.06436: stdout chunk (state=3): >>> # destroy _bz2 # destroy _compression # destroy _lzma<<< 13731 1727203825.06464: stdout chunk (state=3): >>> # destroy _blake2 # destroy binascii # destroy struct # destroy zlib # destroy bz2 # destroy lzma # destroy zipfile._path<<< 13731 1727203825.06500: stdout chunk (state=3): >>> # destroy zipfile # destroy pathlib # destroy zipfile._path.glob # destroy fnmatch<<< 13731 1727203825.06564: stdout chunk (state=3): >>> # destroy ipaddress # destroy ntpath <<< 13731 1727203825.06585: stdout chunk (state=3): >>># destroy importlib # destroy zipimport<<< 13731 1727203825.06618: stdout chunk (state=3): >>> # destroy __main__ # destroy tempfile # destroy systemd.journal # destroy systemd.daemon # destroy ansible.module_utils.compat.selinux # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json<<< 13731 1727203825.06658: stdout chunk (state=3): >>> # destroy grp # destroy encodings # destroy _locale # destroy pwd # destroy locale <<< 13731 1727203825.06691: stdout chunk (state=3): >>># destroy signal # destroy fcntl # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog<<< 13731 1727203825.06756: stdout chunk (state=3): >>> # destroy uuid # destroy selectors # destroy errno <<< 13731 1727203825.06759: stdout chunk (state=3): >>># destroy array # destroy datetime<<< 13731 1727203825.06791: stdout chunk (state=3): >>> # destroy selinux # destroy shutil<<< 13731 1727203825.06809: stdout chunk (state=3): >>> # destroy distro # destroy distro.distro<<< 13731 1727203825.06819: stdout chunk (state=3): >>> # destroy argparse # destroy json # destroy logging # destroy shlex # destroy subprocess<<< 13731 1727203825.06884: stdout chunk (state=3): >>> # cleanup[3] wiping selinux._selinux <<< 13731 1727203825.06918: stdout chunk (state=3): >>># cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # cleanup[3] wiping systemd._daemon<<< 13731 1727203825.06968: stdout chunk (state=3): >>> # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache<<< 13731 1727203825.06983: stdout chunk (state=3): >>> # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform # cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 <<< 13731 1727203825.07018: stdout chunk (state=3): >>># cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external<<< 13731 1727203825.07052: stdout chunk (state=3): >>> # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler<<< 13731 1727203825.07083: stdout chunk (state=3): >>> # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser # cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types # cleanup[3] wiping encodings.utf_8_sig<<< 13731 1727203825.07123: stdout chunk (state=3): >>> # cleanup[3] wiping os # destroy posixpath<<< 13731 1727203825.07137: stdout chunk (state=3): >>> # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time <<< 13731 1727203825.07183: stdout chunk (state=3): >>># cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins<<< 13731 1727203825.07202: stdout chunk (state=3): >>> # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 <<< 13731 1727203825.07293: stdout chunk (state=3): >>># destroy systemd._reader # destroy systemd._journal # destroy _datetime <<< 13731 1727203825.07404: stdout chunk (state=3): >>># destroy sys.monitoring <<< 13731 1727203825.07418: stdout chunk (state=3): >>># destroy _socket <<< 13731 1727203825.07442: stdout chunk (state=3): >>># destroy _collections <<< 13731 1727203825.07505: stdout chunk (state=3): >>># destroy platform # destroy _uuid<<< 13731 1727203825.07530: stdout chunk (state=3): >>> # destroy stat # destroy genericpath # destroy re._parser # destroy tokenize<<< 13731 1727203825.07579: stdout chunk (state=3): >>> # destroy ansible.module_utils.six.moves.urllib # destroy copyreg<<< 13731 1727203825.07595: stdout chunk (state=3): >>> # destroy contextlib <<< 13731 1727203825.07641: stdout chunk (state=3): >>># destroy _typing # destroy _tokenize<<< 13731 1727203825.07665: stdout chunk (state=3): >>> # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser<<< 13731 1727203825.07714: stdout chunk (state=3): >>> # destroy functools # destroy operator # destroy ansible.module_utils.six.moves # destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal<<< 13731 1727203825.07717: stdout chunk (state=3): >>> # clear sys.meta_path # clear sys.modules <<< 13731 1727203825.07798: stdout chunk (state=3): >>># destroy _frozen_importlib <<< 13731 1727203825.07864: stdout chunk (state=3): >>># destroy codecs <<< 13731 1727203825.07904: stdout chunk (state=3): >>># destroy encodings.aliases # destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy _codecs <<< 13731 1727203825.07930: stdout chunk (state=3): >>># destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections<<< 13731 1727203825.07955: stdout chunk (state=3): >>> # destroy threading # destroy atexit # destroy _warnings<<< 13731 1727203825.08004: stdout chunk (state=3): >>> # destroy math # destroy _bisect # destroy time<<< 13731 1727203825.08007: stdout chunk (state=3): >>> # destroy _random # destroy _weakref<<< 13731 1727203825.08043: stdout chunk (state=3): >>> <<< 13731 1727203825.08068: stdout chunk (state=3): >>># destroy _hashlib # destroy _operator # destroy _string # destroy re<<< 13731 1727203825.08098: stdout chunk (state=3): >>> # destroy itertools <<< 13731 1727203825.08135: stdout chunk (state=3): >>># destroy _abc # destroy _sre<<< 13731 1727203825.08138: stdout chunk (state=3): >>> # destroy posix # destroy _functools # destroy builtins # destroy _thread # clear sys.audit hooks<<< 13731 1727203825.08192: stdout chunk (state=3): >>> <<< 13731 1727203825.08791: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203825.08798: stdout chunk (state=3): >>><<< 13731 1727203825.08800: stderr chunk (state=3): >>><<< 13731 1727203825.08810: _low_level_execute_command() done: rc=0, stdout=import _frozen_importlib # frozen import _imp # builtin import '_thread' # import '_warnings' # import '_weakref' # import '_io' # import 'marshal' # import 'posix' # import '_frozen_importlib_external' # # installing zipimport hook import 'time' # import 'zipimport' # # installed zipimport hook # /usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/encodings/__init__.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/__init__.cpython-312.pyc' import '_codecs' # import 'codecs' # # /usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc matches /usr/lib64/python3.12/encodings/aliases.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/aliases.cpython-312.pyc' import 'encodings.aliases' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c340184d0> import 'encodings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33fe7b30> # /usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8.cpython-312.pyc' import 'encodings.utf_8' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3401aa50> import '_signal' # import '_abc' # import 'abc' # import 'io' # import '_stat' # import 'stat' # import '_collections_abc' # import 'genericpath' # import 'posixpath' # import 'os' # import '_sitebuiltins' # Processing user site-packages Processing global site-packages Adding directory: '/usr/local/lib/python3.12/site-packages' Adding directory: '/usr/lib64/python3.12/site-packages' Adding directory: '/usr/lib/python3.12/site-packages' Processing .pth file: '/usr/lib/python3.12/site-packages/distutils-precedence.pth' # /usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc matches /usr/lib64/python3.12/encodings/utf_8_sig.py # code object from '/usr/lib64/python3.12/encodings/__pycache__/utf_8_sig.cpython-312.pyc' import 'encodings.utf_8_sig' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33de9130> # /usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/_distutils_hack/__init__.py # code object from '/usr/lib/python3.12/site-packages/_distutils_hack/__pycache__/__init__.cpython-312.pyc' import '_distutils_hack' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33dea060> import 'site' # Python 3.12.5 (main, Aug 23 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-1)] on linux Type "help", "copyright", "credits" or "license" for more information. # /usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc matches /usr/lib64/python3.12/base64.py # code object from '/usr/lib64/python3.12/__pycache__/base64.cpython-312.pyc' # /usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/re/__init__.py # code object from '/usr/lib64/python3.12/re/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc matches /usr/lib64/python3.12/enum.py # code object from '/usr/lib64/python3.12/__pycache__/enum.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/types.cpython-312.pyc matches /usr/lib64/python3.12/types.py # code object from '/usr/lib64/python3.12/__pycache__/types.cpython-312.pyc' import 'types' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e27f50> # /usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc matches /usr/lib64/python3.12/operator.py # code object from '/usr/lib64/python3.12/__pycache__/operator.cpython-312.pyc' import '_operator' # import 'operator' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e3c0e0> # /usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc matches /usr/lib64/python3.12/functools.py # code object from '/usr/lib64/python3.12/__pycache__/functools.cpython-312.pyc' # /usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/collections/__init__.py # code object from '/usr/lib64/python3.12/collections/__pycache__/__init__.cpython-312.pyc' import 'itertools' # # /usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc matches /usr/lib64/python3.12/keyword.py # code object from '/usr/lib64/python3.12/__pycache__/keyword.cpython-312.pyc' import 'keyword' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e5f980> # /usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc matches /usr/lib64/python3.12/reprlib.py # code object from '/usr/lib64/python3.12/__pycache__/reprlib.cpython-312.pyc' import 'reprlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e5ff50> import '_collections' # import 'collections' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e3fc20> import '_functools' # import 'functools' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e3d340> import 'enum' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e25100> # /usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc matches /usr/lib64/python3.12/re/_compiler.py # code object from '/usr/lib64/python3.12/re/__pycache__/_compiler.cpython-312.pyc' import '_sre' # # /usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc matches /usr/lib64/python3.12/re/_parser.py # code object from '/usr/lib64/python3.12/re/__pycache__/_parser.cpython-312.pyc' # /usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc matches /usr/lib64/python3.12/re/_constants.py # code object from '/usr/lib64/python3.12/re/__pycache__/_constants.cpython-312.pyc' import 're._constants' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e83950> import 're._parser' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e82570> # /usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc matches /usr/lib64/python3.12/re/_casefix.py # code object from '/usr/lib64/python3.12/re/__pycache__/_casefix.cpython-312.pyc' import 're._casefix' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e3e210> import 're._compiler' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e80d70> # /usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc matches /usr/lib64/python3.12/copyreg.py # code object from '/usr/lib64/python3.12/__pycache__/copyreg.cpython-312.pyc' import 'copyreg' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb0950> import 're' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e24380> # /usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc matches /usr/lib64/python3.12/struct.py # code object from '/usr/lib64/python3.12/__pycache__/struct.cpython-312.pyc' # extension module '_struct' loaded from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' # extension module '_struct' executed from '/usr/lib64/python3.12/lib-dynload/_struct.cpython-312-x86_64-linux-gnu.so' import '_struct' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33eb0e00> import 'struct' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb0cb0> # extension module 'binascii' loaded from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' # extension module 'binascii' executed from '/usr/lib64/python3.12/lib-dynload/binascii.cpython-312-x86_64-linux-gnu.so' import 'binascii' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33eb10a0> import 'base64' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33e22ea0> # /usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/importlib/__init__.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc matches /usr/lib64/python3.12/warnings.py # code object from '/usr/lib64/python3.12/__pycache__/warnings.cpython-312.pyc' import 'warnings' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb1760> import 'importlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb1460> import 'importlib.machinery' # # /usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc matches /usr/lib64/python3.12/importlib/_abc.py # code object from '/usr/lib64/python3.12/importlib/__pycache__/_abc.cpython-312.pyc' import 'importlib._abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb2660> import 'importlib.util' # import 'runpy' # # /usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc matches /usr/lib64/python3.12/shutil.py # code object from '/usr/lib64/python3.12/__pycache__/shutil.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc matches /usr/lib64/python3.12/fnmatch.py # code object from '/usr/lib64/python3.12/__pycache__/fnmatch.cpython-312.pyc' import 'fnmatch' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33ecc860> import 'errno' # # extension module 'zlib' loaded from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' # extension module 'zlib' executed from '/usr/lib64/python3.12/lib-dynload/zlib.cpython-312-x86_64-linux-gnu.so' import 'zlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33ecdfa0> # /usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc matches /usr/lib64/python3.12/bz2.py # code object from '/usr/lib64/python3.12/__pycache__/bz2.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc matches /usr/lib64/python3.12/_compression.py # code object from '/usr/lib64/python3.12/__pycache__/_compression.cpython-312.pyc' import '_compression' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33ecee40> # extension module '_bz2' loaded from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' # extension module '_bz2' executed from '/usr/lib64/python3.12/lib-dynload/_bz2.cpython-312-x86_64-linux-gnu.so' import '_bz2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33ecf4a0> import 'bz2' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33ece390> # /usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc matches /usr/lib64/python3.12/lzma.py # code object from '/usr/lib64/python3.12/__pycache__/lzma.cpython-312.pyc' # extension module '_lzma' loaded from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' # extension module '_lzma' executed from '/usr/lib64/python3.12/lib-dynload/_lzma.cpython-312-x86_64-linux-gnu.so' import '_lzma' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33ecff20> import 'lzma' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33ecf650> import 'shutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb2690> # /usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc matches /usr/lib64/python3.12/tempfile.py # code object from '/usr/lib64/python3.12/__pycache__/tempfile.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/random.cpython-312.pyc matches /usr/lib64/python3.12/random.py # code object from '/usr/lib64/python3.12/__pycache__/random.cpython-312.pyc' # extension module 'math' loaded from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' # extension module 'math' executed from '/usr/lib64/python3.12/lib-dynload/math.cpython-312-x86_64-linux-gnu.so' import 'math' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c5bda0> # /usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc matches /usr/lib64/python3.12/bisect.py # code object from '/usr/lib64/python3.12/__pycache__/bisect.cpython-312.pyc' # extension module '_bisect' loaded from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' # extension module '_bisect' executed from '/usr/lib64/python3.12/lib-dynload/_bisect.cpython-312-x86_64-linux-gnu.so' import '_bisect' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c848f0> import 'bisect' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c84650> # extension module '_random' loaded from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' # extension module '_random' executed from '/usr/lib64/python3.12/lib-dynload/_random.cpython-312-x86_64-linux-gnu.so' import '_random' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c84920> # /usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc matches /usr/lib64/python3.12/hashlib.py # code object from '/usr/lib64/python3.12/__pycache__/hashlib.cpython-312.pyc' # extension module '_hashlib' loaded from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' # extension module '_hashlib' executed from '/usr/lib64/python3.12/lib-dynload/_hashlib.cpython-312-x86_64-linux-gnu.so' import '_hashlib' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c85250> # extension module '_blake2' loaded from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' # extension module '_blake2' executed from '/usr/lib64/python3.12/lib-dynload/_blake2.cpython-312-x86_64-linux-gnu.so' import '_blake2' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33c85c40> import 'hashlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c84b00> import 'random' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c59f40> # /usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc matches /usr/lib64/python3.12/weakref.py # code object from '/usr/lib64/python3.12/__pycache__/weakref.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc matches /usr/lib64/python3.12/_weakrefset.py # code object from '/usr/lib64/python3.12/__pycache__/_weakrefset.cpython-312.pyc' import '_weakrefset' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c87050> import 'weakref' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c85d90> import 'tempfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33eb2d80> # /usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/__init__.py # code object from '/usr/lib64/python3.12/zipfile/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc matches /usr/lib64/python3.12/threading.py # code object from '/usr/lib64/python3.12/__pycache__/threading.cpython-312.pyc' import 'threading' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33caf3e0> # /usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/__init__.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc matches /usr/lib64/python3.12/contextlib.py # code object from '/usr/lib64/python3.12/__pycache__/contextlib.cpython-312.pyc' import 'contextlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33cd37a0> # /usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc matches /usr/lib64/python3.12/pathlib.py # code object from '/usr/lib64/python3.12/__pycache__/pathlib.cpython-312.pyc' import 'ntpath' # # /usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/urllib/__init__.py # code object from '/usr/lib64/python3.12/urllib/__pycache__/__init__.cpython-312.pyc' import 'urllib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33d30590> # /usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc matches /usr/lib64/python3.12/urllib/parse.py # code object from '/usr/lib64/python3.12/urllib/__pycache__/parse.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc matches /usr/lib64/python3.12/ipaddress.py # code object from '/usr/lib64/python3.12/__pycache__/ipaddress.cpython-312.pyc' import 'ipaddress' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33d32cf0> import 'urllib.parse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33d306b0> import 'pathlib' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33cf95b0> # /usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc matches /usr/lib64/python3.12/zipfile/_path/glob.py # code object from '/usr/lib64/python3.12/zipfile/_path/__pycache__/glob.cpython-312.pyc' import 'zipfile._path.glob' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b39700> import 'zipfile._path' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33cd25a0> import 'zipfile' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33c87fb0> # code object from '/usr/lib64/python3.12/encodings/cp437.pyc' import 'encodings.cp437' # <_frozen_importlib_external.SourcelessFileLoader object at 0x7f0c33cd2900> # zipimport: found 30 names in '/tmp/ansible_stat_payload_alr9sf9p/ansible_stat_payload.zip' # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc matches /usr/lib64/python3.12/pkgutil.py # code object from '/usr/lib64/python3.12/__pycache__/pkgutil.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc matches /usr/lib64/python3.12/typing.py # code object from '/usr/lib64/python3.12/__pycache__/typing.cpython-312.pyc' # /usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc matches /usr/lib64/python3.12/collections/abc.py # code object from '/usr/lib64/python3.12/collections/__pycache__/abc.cpython-312.pyc' import 'collections.abc' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b8b470> import '_typing' # import 'typing' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b6e360> import 'pkgutil' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b6d4c0> # zipimport: zlib available import 'ansible' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc matches /usr/lib64/python3.12/__future__.py # code object from '/usr/lib64/python3.12/__pycache__/__future__.cpython-312.pyc' import '__future__' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33b89340> # /usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/json/__init__.py # code object from '/usr/lib64/python3.12/json/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc matches /usr/lib64/python3.12/json/decoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/decoder.cpython-312.pyc' # /usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc matches /usr/lib64/python3.12/json/scanner.py # code object from '/usr/lib64/python3.12/json/__pycache__/scanner.cpython-312.pyc' # extension module '_json' loaded from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' # extension module '_json' executed from '/usr/lib64/python3.12/lib-dynload/_json.cpython-312-x86_64-linux-gnu.so' import '_json' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33bb6de0> import 'json.scanner' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bb6b70> import 'json.decoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bb6480> # /usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc matches /usr/lib64/python3.12/json/encoder.py # code object from '/usr/lib64/python3.12/json/__pycache__/encoder.cpython-312.pyc' import 'json.encoder' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bb6ed0> import 'json' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3401a9c0> import 'atexit' # # extension module 'grp' loaded from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' # extension module 'grp' executed from '/usr/lib64/python3.12/lib-dynload/grp.cpython-312-x86_64-linux-gnu.so' import 'grp' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33bb7b00> # extension module 'fcntl' loaded from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' # extension module 'fcntl' executed from '/usr/lib64/python3.12/lib-dynload/fcntl.cpython-312-x86_64-linux-gnu.so' import 'fcntl' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33bb7d40> # /usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc matches /usr/lib64/python3.12/locale.py # code object from '/usr/lib64/python3.12/__pycache__/locale.cpython-312.pyc' import '_locale' # import 'locale' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bdc230> import 'pwd' # # /usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc matches /usr/lib64/python3.12/platform.py # code object from '/usr/lib64/python3.12/__pycache__/platform.cpython-312.pyc' import 'platform' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33511f10> # extension module 'select' loaded from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' # extension module 'select' executed from '/usr/lib64/python3.12/lib-dynload/select.cpython-312-x86_64-linux-gnu.so' import 'select' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33513b30> # /usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc matches /usr/lib64/python3.12/selectors.py # code object from '/usr/lib64/python3.12/__pycache__/selectors.cpython-312.pyc' import 'selectors' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33514500> # /usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc matches /usr/lib64/python3.12/shlex.py # code object from '/usr/lib64/python3.12/__pycache__/shlex.cpython-312.pyc' import 'shlex' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335153d0> # /usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc matches /usr/lib64/python3.12/subprocess.py # code object from '/usr/lib64/python3.12/__pycache__/subprocess.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc matches /usr/lib64/python3.12/signal.py # code object from '/usr/lib64/python3.12/__pycache__/signal.cpython-312.pyc' import 'signal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335200e0> # extension module '_posixsubprocess' loaded from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' # extension module '_posixsubprocess' executed from '/usr/lib64/python3.12/lib-dynload/_posixsubprocess.cpython-312-x86_64-linux-gnu.so' import '_posixsubprocess' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33520230> import 'subprocess' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335163c0> # /usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc matches /usr/lib64/python3.12/traceback.py # code object from '/usr/lib64/python3.12/__pycache__/traceback.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc matches /usr/lib64/python3.12/linecache.py # code object from '/usr/lib64/python3.12/__pycache__/linecache.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc matches /usr/lib64/python3.12/tokenize.py # code object from '/usr/lib64/python3.12/__pycache__/tokenize.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/token.cpython-312.pyc matches /usr/lib64/python3.12/token.py # code object from '/usr/lib64/python3.12/__pycache__/token.cpython-312.pyc' import 'token' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33523f80> import '_tokenize' # import 'tokenize' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33522a50> import 'linecache' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335227b0> # /usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc matches /usr/lib64/python3.12/textwrap.py # code object from '/usr/lib64/python3.12/__pycache__/textwrap.cpython-312.pyc' import 'textwrap' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33522d20> import 'traceback' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335168d0> # extension module 'syslog' loaded from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' # extension module 'syslog' executed from '/usr/lib64/python3.12/lib-dynload/syslog.cpython-312-x86_64-linux-gnu.so' import 'syslog' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c335681a0> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/__init__.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/__init__.cpython-312.pyc' import 'systemd' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33568350> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/journal.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/journal.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc matches /usr/lib64/python3.12/datetime.py # code object from '/usr/lib64/python3.12/__pycache__/datetime.cpython-312.pyc' # extension module '_datetime' loaded from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' # extension module '_datetime' executed from '/usr/lib64/python3.12/lib-dynload/_datetime.cpython-312-x86_64-linux-gnu.so' import '_datetime' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33569df0> import 'datetime' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33569bb0> # /usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc matches /usr/lib64/python3.12/uuid.py # code object from '/usr/lib64/python3.12/__pycache__/uuid.cpython-312.pyc' # extension module '_uuid' loaded from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' # extension module '_uuid' executed from '/usr/lib64/python3.12/lib-dynload/_uuid.cpython-312-x86_64-linux-gnu.so' import '_uuid' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c3356c2f0> import 'uuid' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3356a4e0> # /usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/logging/__init__.py # code object from '/usr/lib64/python3.12/logging/__pycache__/__init__.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/string.cpython-312.pyc matches /usr/lib64/python3.12/string.py # code object from '/usr/lib64/python3.12/__pycache__/string.cpython-312.pyc' import '_string' # import 'string' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3356faa0> import 'logging' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3356c4a0> # extension module 'systemd._journal' loaded from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._journal' executed from '/usr/lib64/python3.12/site-packages/systemd/_journal.cpython-312-x86_64-linux-gnu.so' import 'systemd._journal' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c335705c0> # extension module 'systemd._reader' loaded from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._reader' executed from '/usr/lib64/python3.12/site-packages/systemd/_reader.cpython-312-x86_64-linux-gnu.so' import 'systemd._reader' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33570e90> # extension module 'systemd.id128' loaded from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd.id128' executed from '/usr/lib64/python3.12/site-packages/systemd/id128.cpython-312-x86_64-linux-gnu.so' import 'systemd.id128' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33570920> import 'systemd.journal' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33568470> # /usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/systemd/daemon.py # code object from '/usr/lib64/python3.12/site-packages/systemd/__pycache__/daemon.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc matches /usr/lib64/python3.12/socket.py # code object from '/usr/lib64/python3.12/__pycache__/socket.cpython-312.pyc' # extension module '_socket' loaded from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' # extension module '_socket' executed from '/usr/lib64/python3.12/lib-dynload/_socket.cpython-312-x86_64-linux-gnu.so' import '_socket' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c335fc260> # extension module 'array' loaded from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' # extension module 'array' executed from '/usr/lib64/python3.12/lib-dynload/array.cpython-312-x86_64-linux-gnu.so' import 'array' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c335fd520> import 'socket' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33572a20> # extension module 'systemd._daemon' loaded from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' # extension module 'systemd._daemon' executed from '/usr/lib64/python3.12/site-packages/systemd/_daemon.cpython-312-x86_64-linux-gnu.so' import 'systemd._daemon' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33573da0> import 'systemd.daemon' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33572630> # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.compat' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.text' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.six' # import 'ansible.module_utils.six.moves' # import 'ansible.module_utils.six.moves.collections_abc' # import 'ansible.module_utils.common.text.converters' # # /usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/__init__.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/__init__.cpython-312.pyc' # extension module '_ctypes' loaded from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' # extension module '_ctypes' executed from '/usr/lib64/python3.12/lib-dynload/_ctypes.cpython-312-x86_64-linux-gnu.so' import '_ctypes' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c33401820> # /usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc matches /usr/lib64/python3.12/ctypes/_endian.py # code object from '/usr/lib64/python3.12/ctypes/__pycache__/_endian.cpython-312.pyc' import 'ctypes._endian' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c334027b0> import 'ctypes' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c335fd760> import 'ansible.module_utils.compat.selinux' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils._text' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc matches /usr/lib64/python3.12/copy.py # code object from '/usr/lib64/python3.12/__pycache__/copy.cpython-312.pyc' import 'copy' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33402c60> # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.collections' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.warnings' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.errors' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.parsing' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.parsing.convert_bool' # # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc matches /usr/lib64/python3.12/ast.py # code object from '/usr/lib64/python3.12/__pycache__/ast.cpython-312.pyc' import '_ast' # import 'ast' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c334039e0> # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.text.formatters' # import 'ansible.module_utils.common.validation' # import 'ansible.module_utils.common.parameters' # import 'ansible.module_utils.common.arg_spec' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common.locale' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc matches /usr/lib64/python3.12/site-packages/selinux/__init__.py # code object from '/usr/lib64/python3.12/site-packages/selinux/__pycache__/__init__.cpython-312.pyc' # extension module 'selinux._selinux' loaded from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' # extension module 'selinux._selinux' executed from '/usr/lib64/python3.12/site-packages/selinux/_selinux.cpython-312-x86_64-linux-gnu.so' import 'selinux._selinux' # <_frozen_importlib_external.ExtensionFileLoader object at 0x7f0c3340e3c0> import 'selinux' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3340b650> import 'ansible.module_utils.common.file' # import 'ansible.module_utils.common.process' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available # /usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/__init__.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/__init__.cpython-312.pyc' # /usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc matches /usr/lib/python3.12/site-packages/distro/distro.py # code object from '/usr/lib/python3.12/site-packages/distro/__pycache__/distro.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc matches /usr/lib64/python3.12/argparse.py # code object from '/usr/lib64/python3.12/__pycache__/argparse.cpython-312.pyc' # /usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc matches /usr/lib64/python3.12/gettext.py # code object from '/usr/lib64/python3.12/__pycache__/gettext.cpython-312.pyc' import 'gettext' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33be2ba0> import 'argparse' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c33bfe870> import 'distro.distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c3340e480> import 'distro' # <_frozen_importlib_external.SourceFileLoader object at 0x7f0c334031d0> # destroy ansible.module_utils.distro import 'ansible.module_utils.distro' # # zipimport: zlib available # zipimport: zlib available import 'ansible.module_utils.common._utils' # import 'ansible.module_utils.common.sys_info' # import 'ansible.module_utils.basic' # # zipimport: zlib available # zipimport: zlib available import 'ansible.modules' # # zipimport: zlib available # zipimport: zlib available # zipimport: zlib available {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/run/ostree-booted", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} # destroy __main__ # clear sys.path_importer_cache # clear sys.path_hooks # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_exc # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # destroy base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # destroy socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # destroy _sitebuiltins # destroy importlib.machinery # destroy importlib._abc # destroy importlib.util # destroy _bz2 # destroy _compression # destroy _lzma # destroy _blake2 # destroy binascii # destroy struct # destroy zlib # destroy bz2 # destroy lzma # destroy zipfile._path # destroy zipfile # destroy pathlib # destroy zipfile._path.glob # destroy fnmatch # destroy ipaddress # destroy ntpath # destroy importlib # destroy zipimport # destroy __main__ # destroy tempfile # destroy systemd.journal # destroy systemd.daemon # destroy ansible.module_utils.compat.selinux # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy grp # destroy encodings # destroy _locale # destroy pwd # destroy locale # destroy signal # destroy fcntl # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog # destroy uuid # destroy selectors # destroy errno # destroy array # destroy datetime # destroy selinux # destroy shutil # destroy distro # destroy distro.distro # destroy argparse # destroy json # destroy logging # destroy shlex # destroy subprocess # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform # cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser # cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types # cleanup[3] wiping encodings.utf_8_sig # cleanup[3] wiping os # destroy posixpath # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy sys.monitoring # destroy _socket # destroy _collections # destroy platform # destroy _uuid # destroy stat # destroy genericpath # destroy re._parser # destroy tokenize # destroy ansible.module_utils.six.moves.urllib # destroy copyreg # destroy contextlib # destroy _typing # destroy _tokenize # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy operator # destroy ansible.module_utils.six.moves # destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal # clear sys.meta_path # clear sys.modules # destroy _frozen_importlib # destroy codecs # destroy encodings.aliases # destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy _codecs # destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections # destroy threading # destroy atexit # destroy _warnings # destroy math # destroy _bisect # destroy time # destroy _random # destroy _weakref # destroy _hashlib # destroy _operator # destroy _string # destroy re # destroy itertools # destroy _abc # destroy _sre # destroy posix # destroy _functools # destroy builtins # destroy _thread # clear sys.audit hooks , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. [WARNING]: Module invocation had junk after the JSON data: # destroy __main__ # clear sys.path_importer_cache # clear sys.path_hooks # clear builtins._ # clear sys.path # clear sys.argv # clear sys.ps1 # clear sys.ps2 # clear sys.last_exc # clear sys.last_type # clear sys.last_value # clear sys.last_traceback # clear sys.__interactivehook__ # clear sys.meta_path # restore sys.stdin # restore sys.stdout # restore sys.stderr # cleanup[2] removing sys # cleanup[2] removing builtins # cleanup[2] removing _frozen_importlib # cleanup[2] removing _imp # cleanup[2] removing _thread # cleanup[2] removing _warnings # cleanup[2] removing _weakref # cleanup[2] removing _io # cleanup[2] removing marshal # cleanup[2] removing posix # cleanup[2] removing _frozen_importlib_external # cleanup[2] removing time # cleanup[2] removing zipimport # cleanup[2] removing _codecs # cleanup[2] removing codecs # cleanup[2] removing encodings.aliases # cleanup[2] removing encodings # cleanup[2] removing encodings.utf_8 # cleanup[2] removing _signal # cleanup[2] removing _abc # cleanup[2] removing abc # cleanup[2] removing io # cleanup[2] removing __main__ # cleanup[2] removing _stat # cleanup[2] removing stat # cleanup[2] removing _collections_abc # cleanup[2] removing genericpath # cleanup[2] removing posixpath # cleanup[2] removing os.path # cleanup[2] removing os # cleanup[2] removing _sitebuiltins # cleanup[2] removing encodings.utf_8_sig # cleanup[2] removing _distutils_hack # destroy _distutils_hack # cleanup[2] removing site # destroy site # cleanup[2] removing types # cleanup[2] removing _operator # cleanup[2] removing operator # cleanup[2] removing itertools # cleanup[2] removing keyword # destroy keyword # cleanup[2] removing reprlib # destroy reprlib # cleanup[2] removing _collections # cleanup[2] removing collections # cleanup[2] removing _functools # cleanup[2] removing functools # cleanup[2] removing enum # cleanup[2] removing _sre # cleanup[2] removing re._constants # cleanup[2] removing re._parser # cleanup[2] removing re._casefix # cleanup[2] removing re._compiler # cleanup[2] removing copyreg # cleanup[2] removing re # cleanup[2] removing _struct # cleanup[2] removing struct # cleanup[2] removing binascii # cleanup[2] removing base64 # destroy base64 # cleanup[2] removing importlib._bootstrap # cleanup[2] removing importlib._bootstrap_external # cleanup[2] removing warnings # cleanup[2] removing importlib # cleanup[2] removing importlib.machinery # cleanup[2] removing importlib._abc # cleanup[2] removing importlib.util # cleanup[2] removing runpy # destroy runpy # cleanup[2] removing fnmatch # cleanup[2] removing errno # cleanup[2] removing zlib # cleanup[2] removing _compression # cleanup[2] removing _bz2 # cleanup[2] removing bz2 # cleanup[2] removing _lzma # cleanup[2] removing lzma # cleanup[2] removing shutil # cleanup[2] removing math # cleanup[2] removing _bisect # cleanup[2] removing bisect # destroy bisect # cleanup[2] removing _random # cleanup[2] removing _hashlib # cleanup[2] removing _blake2 # cleanup[2] removing hashlib # cleanup[2] removing random # destroy random # cleanup[2] removing _weakrefset # destroy _weakrefset # cleanup[2] removing weakref # cleanup[2] removing tempfile # cleanup[2] removing threading # cleanup[2] removing contextlib # cleanup[2] removing ntpath # cleanup[2] removing urllib # destroy urllib # cleanup[2] removing ipaddress # cleanup[2] removing urllib.parse # destroy urllib.parse # cleanup[2] removing pathlib # cleanup[2] removing zipfile._path.glob # cleanup[2] removing zipfile._path # cleanup[2] removing zipfile # cleanup[2] removing encodings.cp437 # cleanup[2] removing collections.abc # cleanup[2] removing _typing # cleanup[2] removing typing # destroy typing # cleanup[2] removing pkgutil # destroy pkgutil # cleanup[2] removing ansible # destroy ansible # cleanup[2] removing ansible.module_utils # destroy ansible.module_utils # cleanup[2] removing __future__ # destroy __future__ # cleanup[2] removing _json # cleanup[2] removing json.scanner # cleanup[2] removing json.decoder # cleanup[2] removing json.encoder # cleanup[2] removing json # cleanup[2] removing atexit # cleanup[2] removing grp # cleanup[2] removing fcntl # cleanup[2] removing _locale # cleanup[2] removing locale # cleanup[2] removing pwd # cleanup[2] removing platform # cleanup[2] removing select # cleanup[2] removing selectors # cleanup[2] removing shlex # cleanup[2] removing signal # cleanup[2] removing _posixsubprocess # cleanup[2] removing subprocess # cleanup[2] removing token # destroy token # cleanup[2] removing _tokenize # cleanup[2] removing tokenize # cleanup[2] removing linecache # cleanup[2] removing textwrap # cleanup[2] removing traceback # cleanup[2] removing syslog # cleanup[2] removing systemd # destroy systemd # cleanup[2] removing _datetime # cleanup[2] removing datetime # cleanup[2] removing _uuid # cleanup[2] removing uuid # cleanup[2] removing _string # cleanup[2] removing string # destroy string # cleanup[2] removing logging # cleanup[2] removing systemd._journal # cleanup[2] removing systemd._reader # cleanup[2] removing systemd.id128 # cleanup[2] removing systemd.journal # cleanup[2] removing _socket # cleanup[2] removing array # cleanup[2] removing socket # destroy socket # cleanup[2] removing systemd._daemon # cleanup[2] removing systemd.daemon # cleanup[2] removing ansible.module_utils.compat # destroy ansible.module_utils.compat # cleanup[2] removing ansible.module_utils.common # destroy ansible.module_utils.common # cleanup[2] removing ansible.module_utils.common.text # destroy ansible.module_utils.common.text # cleanup[2] removing ansible.module_utils.six # destroy ansible.module_utils.six # cleanup[2] removing ansible.module_utils.six.moves # cleanup[2] removing ansible.module_utils.six.moves.collections_abc # cleanup[2] removing ansible.module_utils.common.text.converters # destroy ansible.module_utils.common.text.converters # cleanup[2] removing _ctypes # cleanup[2] removing ctypes._endian # cleanup[2] removing ctypes # destroy ctypes # cleanup[2] removing ansible.module_utils.compat.selinux # cleanup[2] removing ansible.module_utils._text # destroy ansible.module_utils._text # cleanup[2] removing copy # destroy copy # cleanup[2] removing ansible.module_utils.common.collections # destroy ansible.module_utils.common.collections # cleanup[2] removing ansible.module_utils.common.warnings # destroy ansible.module_utils.common.warnings # cleanup[2] removing ansible.module_utils.errors # destroy ansible.module_utils.errors # cleanup[2] removing ansible.module_utils.parsing # destroy ansible.module_utils.parsing # cleanup[2] removing ansible.module_utils.parsing.convert_bool # destroy ansible.module_utils.parsing.convert_bool # cleanup[2] removing _ast # destroy _ast # cleanup[2] removing ast # destroy ast # cleanup[2] removing ansible.module_utils.common.text.formatters # destroy ansible.module_utils.common.text.formatters # cleanup[2] removing ansible.module_utils.common.validation # destroy ansible.module_utils.common.validation # cleanup[2] removing ansible.module_utils.common.parameters # destroy ansible.module_utils.common.parameters # cleanup[2] removing ansible.module_utils.common.arg_spec # destroy ansible.module_utils.common.arg_spec # cleanup[2] removing ansible.module_utils.common.locale # destroy ansible.module_utils.common.locale # cleanup[2] removing swig_runtime_data4 # destroy swig_runtime_data4 # cleanup[2] removing selinux._selinux # cleanup[2] removing selinux # cleanup[2] removing ansible.module_utils.common.file # destroy ansible.module_utils.common.file # cleanup[2] removing ansible.module_utils.common.process # destroy ansible.module_utils.common.process # cleanup[2] removing gettext # destroy gettext # cleanup[2] removing argparse # cleanup[2] removing distro.distro # cleanup[2] removing distro # cleanup[2] removing ansible.module_utils.distro # cleanup[2] removing ansible.module_utils.common._utils # destroy ansible.module_utils.common._utils # cleanup[2] removing ansible.module_utils.common.sys_info # destroy ansible.module_utils.common.sys_info # cleanup[2] removing ansible.module_utils.basic # destroy ansible.module_utils.basic # cleanup[2] removing ansible.modules # destroy ansible.modules # destroy _sitebuiltins # destroy importlib.machinery # destroy importlib._abc # destroy importlib.util # destroy _bz2 # destroy _compression # destroy _lzma # destroy _blake2 # destroy binascii # destroy struct # destroy zlib # destroy bz2 # destroy lzma # destroy zipfile._path # destroy zipfile # destroy pathlib # destroy zipfile._path.glob # destroy fnmatch # destroy ipaddress # destroy ntpath # destroy importlib # destroy zipimport # destroy __main__ # destroy tempfile # destroy systemd.journal # destroy systemd.daemon # destroy ansible.module_utils.compat.selinux # destroy hashlib # destroy json.decoder # destroy json.encoder # destroy json.scanner # destroy _json # destroy grp # destroy encodings # destroy _locale # destroy pwd # destroy locale # destroy signal # destroy fcntl # destroy select # destroy _signal # destroy _posixsubprocess # destroy syslog # destroy uuid # destroy selectors # destroy errno # destroy array # destroy datetime # destroy selinux # destroy shutil # destroy distro # destroy distro.distro # destroy argparse # destroy json # destroy logging # destroy shlex # destroy subprocess # cleanup[3] wiping selinux._selinux # cleanup[3] wiping ctypes._endian # cleanup[3] wiping _ctypes # cleanup[3] wiping ansible.module_utils.six.moves.collections_abc # cleanup[3] wiping ansible.module_utils.six.moves # cleanup[3] wiping systemd._daemon # cleanup[3] wiping _socket # cleanup[3] wiping systemd.id128 # cleanup[3] wiping systemd._reader # cleanup[3] wiping systemd._journal # cleanup[3] wiping _string # cleanup[3] wiping _uuid # cleanup[3] wiping _datetime # cleanup[3] wiping traceback # destroy linecache # destroy textwrap # cleanup[3] wiping tokenize # cleanup[3] wiping _tokenize # cleanup[3] wiping platform # cleanup[3] wiping atexit # cleanup[3] wiping _typing # cleanup[3] wiping collections.abc # cleanup[3] wiping encodings.cp437 # cleanup[3] wiping contextlib # cleanup[3] wiping threading # cleanup[3] wiping weakref # cleanup[3] wiping _hashlib # cleanup[3] wiping _random # cleanup[3] wiping _bisect # cleanup[3] wiping math # cleanup[3] wiping warnings # cleanup[3] wiping importlib._bootstrap_external # cleanup[3] wiping importlib._bootstrap # cleanup[3] wiping _struct # cleanup[3] wiping re # destroy re._constants # destroy re._casefix # destroy re._compiler # destroy enum # cleanup[3] wiping copyreg # cleanup[3] wiping re._parser # cleanup[3] wiping _sre # cleanup[3] wiping functools # cleanup[3] wiping _functools # cleanup[3] wiping collections # destroy _collections_abc # destroy collections.abc # cleanup[3] wiping _collections # cleanup[3] wiping itertools # cleanup[3] wiping operator # cleanup[3] wiping _operator # cleanup[3] wiping types # cleanup[3] wiping encodings.utf_8_sig # cleanup[3] wiping os # destroy posixpath # cleanup[3] wiping genericpath # cleanup[3] wiping stat # cleanup[3] wiping _stat # destroy _stat # cleanup[3] wiping io # destroy abc # cleanup[3] wiping _abc # cleanup[3] wiping encodings.utf_8 # cleanup[3] wiping encodings.aliases # cleanup[3] wiping codecs # cleanup[3] wiping _codecs # cleanup[3] wiping time # cleanup[3] wiping _frozen_importlib_external # cleanup[3] wiping posix # cleanup[3] wiping marshal # cleanup[3] wiping _io # cleanup[3] wiping _weakref # cleanup[3] wiping _warnings # cleanup[3] wiping _thread # cleanup[3] wiping _imp # cleanup[3] wiping _frozen_importlib # cleanup[3] wiping sys # cleanup[3] wiping builtins # destroy selinux._selinux # destroy systemd._daemon # destroy systemd.id128 # destroy systemd._reader # destroy systemd._journal # destroy _datetime # destroy sys.monitoring # destroy _socket # destroy _collections # destroy platform # destroy _uuid # destroy stat # destroy genericpath # destroy re._parser # destroy tokenize # destroy ansible.module_utils.six.moves.urllib # destroy copyreg # destroy contextlib # destroy _typing # destroy _tokenize # destroy ansible.module_utils.six.moves.urllib_parse # destroy ansible.module_utils.six.moves.urllib.error # destroy ansible.module_utils.six.moves.urllib.request # destroy ansible.module_utils.six.moves.urllib.response # destroy ansible.module_utils.six.moves.urllib.robotparser # destroy functools # destroy operator # destroy ansible.module_utils.six.moves # destroy _frozen_importlib_external # destroy _imp # destroy _io # destroy marshal # clear sys.meta_path # clear sys.modules # destroy _frozen_importlib # destroy codecs # destroy encodings.aliases # destroy encodings.utf_8 # destroy encodings.utf_8_sig # destroy encodings.cp437 # destroy _codecs # destroy io # destroy traceback # destroy warnings # destroy weakref # destroy collections # destroy threading # destroy atexit # destroy _warnings # destroy math # destroy _bisect # destroy time # destroy _random # destroy _weakref # destroy _hashlib # destroy _operator # destroy _string # destroy re # destroy itertools # destroy _abc # destroy _sre # destroy posix # destroy _functools # destroy builtins # destroy _thread # clear sys.audit hooks 13731 1727203825.09522: done with _execute_module (stat, {'path': '/run/ostree-booted', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203825.09525: _low_level_execute_command(): starting 13731 1727203825.09528: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203824.5319855-13899-259015273838329/ > /dev/null 2>&1 && sleep 0' 13731 1727203825.09908: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203825.09991: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.10028: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203825.10050: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203825.10067: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203825.10139: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203825.12725: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203825.12730: stdout chunk (state=3): >>><<< 13731 1727203825.12736: stderr chunk (state=3): >>><<< 13731 1727203825.12766: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203825.12772: handler run complete 13731 1727203825.12880: attempt loop complete, returning result 13731 1727203825.12883: _execute() done 13731 1727203825.12885: dumping result to json 13731 1727203825.12887: done dumping result, returning 13731 1727203825.12889: done running TaskExecutor() for managed-node3/TASK: Check if system is ostree [028d2410-947f-82dc-c122-000000000028] 13731 1727203825.12891: sending task result for task 028d2410-947f-82dc-c122-000000000028 13731 1727203825.12947: done sending task result for task 028d2410-947f-82dc-c122-000000000028 13731 1727203825.12949: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } 13731 1727203825.13026: no more pending results, returning what we have 13731 1727203825.13029: results queue empty 13731 1727203825.13030: checking for any_errors_fatal 13731 1727203825.13035: done checking for any_errors_fatal 13731 1727203825.13036: checking for max_fail_percentage 13731 1727203825.13037: done checking for max_fail_percentage 13731 1727203825.13038: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.13039: done checking to see if all hosts have failed 13731 1727203825.13039: getting the remaining hosts for this loop 13731 1727203825.13041: done getting the remaining hosts for this loop 13731 1727203825.13045: getting the next task for host managed-node3 13731 1727203825.13051: done getting next task for host managed-node3 13731 1727203825.13054: ^ task is: TASK: Set flag to indicate system is ostree 13731 1727203825.13057: ^ state is: HOST STATE: block=2, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.13060: getting variables 13731 1727203825.13062: in VariableManager get_vars() 13731 1727203825.13095: Calling all_inventory to load vars for managed-node3 13731 1727203825.13098: Calling groups_inventory to load vars for managed-node3 13731 1727203825.13101: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.13112: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.13114: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.13117: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.13522: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.13733: done with get_vars() 13731 1727203825.13744: done getting variables 13731 1727203825.13845: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [Set flag to indicate system is ostree] *********************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:22 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.671) 0:00:03.369 ***** 13731 1727203825.13879: entering _queue_task() for managed-node3/set_fact 13731 1727203825.13881: Creating lock for set_fact 13731 1727203825.14278: worker is 1 (out of 1 available) 13731 1727203825.14288: exiting _queue_task() for managed-node3/set_fact 13731 1727203825.14297: done queuing things up, now waiting for results queue to drain 13731 1727203825.14299: waiting for pending results... 13731 1727203825.14433: running TaskExecutor() for managed-node3/TASK: Set flag to indicate system is ostree 13731 1727203825.14585: in run() - task 028d2410-947f-82dc-c122-000000000029 13731 1727203825.14589: variable 'ansible_search_path' from source: unknown 13731 1727203825.14592: variable 'ansible_search_path' from source: unknown 13731 1727203825.14606: calling self._execute() 13731 1727203825.14682: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.14700: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.14714: variable 'omit' from source: magic vars 13731 1727203825.15199: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203825.15564: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203825.15568: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203825.15608: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203825.15644: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203825.15738: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203825.15766: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203825.15804: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203825.15837: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203825.15962: Evaluated conditional (not __network_is_ostree is defined): True 13731 1727203825.15999: variable 'omit' from source: magic vars 13731 1727203825.16019: variable 'omit' from source: magic vars 13731 1727203825.16146: variable '__ostree_booted_stat' from source: set_fact 13731 1727203825.16217: variable 'omit' from source: magic vars 13731 1727203825.16233: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203825.16268: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203825.16325: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203825.16328: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203825.16330: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203825.16356: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203825.16365: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.16378: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.16488: Set connection var ansible_pipelining to False 13731 1727203825.16543: Set connection var ansible_shell_type to sh 13731 1727203825.16546: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203825.16548: Set connection var ansible_connection to ssh 13731 1727203825.16550: Set connection var ansible_shell_executable to /bin/sh 13731 1727203825.16552: Set connection var ansible_timeout to 10 13731 1727203825.16554: variable 'ansible_shell_executable' from source: unknown 13731 1727203825.16559: variable 'ansible_connection' from source: unknown 13731 1727203825.16566: variable 'ansible_module_compression' from source: unknown 13731 1727203825.16572: variable 'ansible_shell_type' from source: unknown 13731 1727203825.16579: variable 'ansible_shell_executable' from source: unknown 13731 1727203825.16587: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.16596: variable 'ansible_pipelining' from source: unknown 13731 1727203825.16602: variable 'ansible_timeout' from source: unknown 13731 1727203825.16653: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.16723: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203825.16736: variable 'omit' from source: magic vars 13731 1727203825.16745: starting attempt loop 13731 1727203825.16751: running the handler 13731 1727203825.16778: handler run complete 13731 1727203825.16793: attempt loop complete, returning result 13731 1727203825.16799: _execute() done 13731 1727203825.16871: dumping result to json 13731 1727203825.16874: done dumping result, returning 13731 1727203825.16878: done running TaskExecutor() for managed-node3/TASK: Set flag to indicate system is ostree [028d2410-947f-82dc-c122-000000000029] 13731 1727203825.16880: sending task result for task 028d2410-947f-82dc-c122-000000000029 13731 1727203825.16942: done sending task result for task 028d2410-947f-82dc-c122-000000000029 13731 1727203825.16944: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "__network_is_ostree": false }, "changed": false } 13731 1727203825.17026: no more pending results, returning what we have 13731 1727203825.17029: results queue empty 13731 1727203825.17030: checking for any_errors_fatal 13731 1727203825.17038: done checking for any_errors_fatal 13731 1727203825.17039: checking for max_fail_percentage 13731 1727203825.17041: done checking for max_fail_percentage 13731 1727203825.17042: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.17042: done checking to see if all hosts have failed 13731 1727203825.17043: getting the remaining hosts for this loop 13731 1727203825.17045: done getting the remaining hosts for this loop 13731 1727203825.17049: getting the next task for host managed-node3 13731 1727203825.17058: done getting next task for host managed-node3 13731 1727203825.17061: ^ task is: TASK: Fix CentOS6 Base repo 13731 1727203825.17063: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.17067: getting variables 13731 1727203825.17069: in VariableManager get_vars() 13731 1727203825.17206: Calling all_inventory to load vars for managed-node3 13731 1727203825.17209: Calling groups_inventory to load vars for managed-node3 13731 1727203825.17213: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.17223: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.17226: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.17236: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.17618: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.17813: done with get_vars() 13731 1727203825.17823: done getting variables 13731 1727203825.17948: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [Fix CentOS6 Base repo] *************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:26 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.040) 0:00:03.410 ***** 13731 1727203825.17981: entering _queue_task() for managed-node3/copy 13731 1727203825.18254: worker is 1 (out of 1 available) 13731 1727203825.18269: exiting _queue_task() for managed-node3/copy 13731 1727203825.18283: done queuing things up, now waiting for results queue to drain 13731 1727203825.18285: waiting for pending results... 13731 1727203825.18549: running TaskExecutor() for managed-node3/TASK: Fix CentOS6 Base repo 13731 1727203825.18597: in run() - task 028d2410-947f-82dc-c122-00000000002b 13731 1727203825.18608: variable 'ansible_search_path' from source: unknown 13731 1727203825.18611: variable 'ansible_search_path' from source: unknown 13731 1727203825.18644: calling self._execute() 13731 1727203825.18722: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.18777: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.18780: variable 'omit' from source: magic vars 13731 1727203825.19225: variable 'ansible_distribution' from source: facts 13731 1727203825.19243: Evaluated conditional (ansible_distribution == 'CentOS'): True 13731 1727203825.19362: variable 'ansible_distribution_major_version' from source: facts 13731 1727203825.19386: Evaluated conditional (ansible_distribution_major_version == '6'): False 13731 1727203825.19390: when evaluation is False, skipping this task 13731 1727203825.19392: _execute() done 13731 1727203825.19395: dumping result to json 13731 1727203825.19397: done dumping result, returning 13731 1727203825.19399: done running TaskExecutor() for managed-node3/TASK: Fix CentOS6 Base repo [028d2410-947f-82dc-c122-00000000002b] 13731 1727203825.19401: sending task result for task 028d2410-947f-82dc-c122-00000000002b 13731 1727203825.19518: done sending task result for task 028d2410-947f-82dc-c122-00000000002b 13731 1727203825.19521: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version == '6'", "skip_reason": "Conditional result was False" } 13731 1727203825.19594: no more pending results, returning what we have 13731 1727203825.19598: results queue empty 13731 1727203825.19599: checking for any_errors_fatal 13731 1727203825.19604: done checking for any_errors_fatal 13731 1727203825.19605: checking for max_fail_percentage 13731 1727203825.19607: done checking for max_fail_percentage 13731 1727203825.19607: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.19608: done checking to see if all hosts have failed 13731 1727203825.19609: getting the remaining hosts for this loop 13731 1727203825.19611: done getting the remaining hosts for this loop 13731 1727203825.19614: getting the next task for host managed-node3 13731 1727203825.19621: done getting next task for host managed-node3 13731 1727203825.19626: ^ task is: TASK: Include the task 'enable_epel.yml' 13731 1727203825.19629: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.19634: getting variables 13731 1727203825.19636: in VariableManager get_vars() 13731 1727203825.19666: Calling all_inventory to load vars for managed-node3 13731 1727203825.19670: Calling groups_inventory to load vars for managed-node3 13731 1727203825.19674: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.19687: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.19690: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.19693: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.19994: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.20193: done with get_vars() 13731 1727203825.20202: done getting variables TASK [Include the task 'enable_epel.yml'] ************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:51 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.023) 0:00:03.433 ***** 13731 1727203825.20292: entering _queue_task() for managed-node3/include_tasks 13731 1727203825.20525: worker is 1 (out of 1 available) 13731 1727203825.20538: exiting _queue_task() for managed-node3/include_tasks 13731 1727203825.20549: done queuing things up, now waiting for results queue to drain 13731 1727203825.20551: waiting for pending results... 13731 1727203825.20793: running TaskExecutor() for managed-node3/TASK: Include the task 'enable_epel.yml' 13731 1727203825.20892: in run() - task 028d2410-947f-82dc-c122-00000000002c 13731 1727203825.20901: variable 'ansible_search_path' from source: unknown 13731 1727203825.20905: variable 'ansible_search_path' from source: unknown 13731 1727203825.20932: calling self._execute() 13731 1727203825.20989: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.20999: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.21002: variable 'omit' from source: magic vars 13731 1727203825.21399: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203825.23181: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203825.23185: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203825.23188: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203825.23190: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203825.23200: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203825.23271: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203825.23306: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203825.23322: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203825.23364: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203825.23421: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203825.23478: variable '__network_is_ostree' from source: set_fact 13731 1727203825.23494: Evaluated conditional (not __network_is_ostree | d(false)): True 13731 1727203825.23500: _execute() done 13731 1727203825.23502: dumping result to json 13731 1727203825.23505: done dumping result, returning 13731 1727203825.23511: done running TaskExecutor() for managed-node3/TASK: Include the task 'enable_epel.yml' [028d2410-947f-82dc-c122-00000000002c] 13731 1727203825.23532: sending task result for task 028d2410-947f-82dc-c122-00000000002c 13731 1727203825.23640: done sending task result for task 028d2410-947f-82dc-c122-00000000002c 13731 1727203825.23643: WORKER PROCESS EXITING 13731 1727203825.23669: no more pending results, returning what we have 13731 1727203825.23674: in VariableManager get_vars() 13731 1727203825.23706: Calling all_inventory to load vars for managed-node3 13731 1727203825.23709: Calling groups_inventory to load vars for managed-node3 13731 1727203825.23713: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.23722: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.23724: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.23727: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.24013: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.24166: done with get_vars() 13731 1727203825.24172: variable 'ansible_search_path' from source: unknown 13731 1727203825.24172: variable 'ansible_search_path' from source: unknown 13731 1727203825.24200: we have included files to process 13731 1727203825.24201: generating all_blocks data 13731 1727203825.24202: done generating all_blocks data 13731 1727203825.24205: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml 13731 1727203825.24206: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml 13731 1727203825.24207: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml 13731 1727203825.24663: done processing included file 13731 1727203825.24665: iterating over new_blocks loaded from include file 13731 1727203825.24665: in VariableManager get_vars() 13731 1727203825.24673: done with get_vars() 13731 1727203825.24674: filtering new block on tags 13731 1727203825.24692: done filtering new block on tags 13731 1727203825.24694: in VariableManager get_vars() 13731 1727203825.24701: done with get_vars() 13731 1727203825.24702: filtering new block on tags 13731 1727203825.24709: done filtering new block on tags 13731 1727203825.24710: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml for managed-node3 13731 1727203825.24714: extending task lists for all hosts with included blocks 13731 1727203825.24769: done extending task lists 13731 1727203825.24770: done processing included files 13731 1727203825.24771: results queue empty 13731 1727203825.24771: checking for any_errors_fatal 13731 1727203825.24773: done checking for any_errors_fatal 13731 1727203825.24774: checking for max_fail_percentage 13731 1727203825.24774: done checking for max_fail_percentage 13731 1727203825.24775: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.24777: done checking to see if all hosts have failed 13731 1727203825.24777: getting the remaining hosts for this loop 13731 1727203825.24778: done getting the remaining hosts for this loop 13731 1727203825.24779: getting the next task for host managed-node3 13731 1727203825.24782: done getting next task for host managed-node3 13731 1727203825.24784: ^ task is: TASK: Create EPEL {{ ansible_distribution_major_version }} 13731 1727203825.24786: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.24788: getting variables 13731 1727203825.24789: in VariableManager get_vars() 13731 1727203825.24796: Calling all_inventory to load vars for managed-node3 13731 1727203825.24797: Calling groups_inventory to load vars for managed-node3 13731 1727203825.24798: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.24802: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.24807: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.24809: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.24904: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.25016: done with get_vars() 13731 1727203825.25022: done getting variables 13731 1727203825.25067: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) 13731 1727203825.25198: variable 'ansible_distribution_major_version' from source: facts TASK [Create EPEL 10] ********************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:8 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.049) 0:00:03.483 ***** 13731 1727203825.25230: entering _queue_task() for managed-node3/command 13731 1727203825.25232: Creating lock for command 13731 1727203825.25429: worker is 1 (out of 1 available) 13731 1727203825.25442: exiting _queue_task() for managed-node3/command 13731 1727203825.25453: done queuing things up, now waiting for results queue to drain 13731 1727203825.25454: waiting for pending results... 13731 1727203825.25595: running TaskExecutor() for managed-node3/TASK: Create EPEL 10 13731 1727203825.25653: in run() - task 028d2410-947f-82dc-c122-000000000046 13731 1727203825.25666: variable 'ansible_search_path' from source: unknown 13731 1727203825.25669: variable 'ansible_search_path' from source: unknown 13731 1727203825.25702: calling self._execute() 13731 1727203825.25754: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.25757: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.25768: variable 'omit' from source: magic vars 13731 1727203825.26036: variable 'ansible_distribution' from source: facts 13731 1727203825.26044: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13731 1727203825.26147: variable 'ansible_distribution_major_version' from source: facts 13731 1727203825.26151: Evaluated conditional (ansible_distribution_major_version in ['7', '8']): False 13731 1727203825.26153: when evaluation is False, skipping this task 13731 1727203825.26158: _execute() done 13731 1727203825.26163: dumping result to json 13731 1727203825.26187: done dumping result, returning 13731 1727203825.26262: done running TaskExecutor() for managed-node3/TASK: Create EPEL 10 [028d2410-947f-82dc-c122-000000000046] 13731 1727203825.26265: sending task result for task 028d2410-947f-82dc-c122-000000000046 13731 1727203825.26327: done sending task result for task 028d2410-947f-82dc-c122-000000000046 13731 1727203825.26330: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version in ['7', '8']", "skip_reason": "Conditional result was False" } 13731 1727203825.26393: no more pending results, returning what we have 13731 1727203825.26396: results queue empty 13731 1727203825.26396: checking for any_errors_fatal 13731 1727203825.26398: done checking for any_errors_fatal 13731 1727203825.26398: checking for max_fail_percentage 13731 1727203825.26400: done checking for max_fail_percentage 13731 1727203825.26400: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.26401: done checking to see if all hosts have failed 13731 1727203825.26401: getting the remaining hosts for this loop 13731 1727203825.26403: done getting the remaining hosts for this loop 13731 1727203825.26406: getting the next task for host managed-node3 13731 1727203825.26410: done getting next task for host managed-node3 13731 1727203825.26412: ^ task is: TASK: Install yum-utils package 13731 1727203825.26415: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.26418: getting variables 13731 1727203825.26419: in VariableManager get_vars() 13731 1727203825.26444: Calling all_inventory to load vars for managed-node3 13731 1727203825.26446: Calling groups_inventory to load vars for managed-node3 13731 1727203825.26449: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.26457: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.26459: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.26462: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.26817: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.27025: done with get_vars() 13731 1727203825.27034: done getting variables 13731 1727203825.27135: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [Install yum-utils package] *********************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:26 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.019) 0:00:03.502 ***** 13731 1727203825.27160: entering _queue_task() for managed-node3/package 13731 1727203825.27161: Creating lock for package 13731 1727203825.27413: worker is 1 (out of 1 available) 13731 1727203825.27425: exiting _queue_task() for managed-node3/package 13731 1727203825.27435: done queuing things up, now waiting for results queue to drain 13731 1727203825.27436: waiting for pending results... 13731 1727203825.27653: running TaskExecutor() for managed-node3/TASK: Install yum-utils package 13731 1727203825.27731: in run() - task 028d2410-947f-82dc-c122-000000000047 13731 1727203825.27740: variable 'ansible_search_path' from source: unknown 13731 1727203825.27743: variable 'ansible_search_path' from source: unknown 13731 1727203825.27782: calling self._execute() 13731 1727203825.27844: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.27849: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.27857: variable 'omit' from source: magic vars 13731 1727203825.28140: variable 'ansible_distribution' from source: facts 13731 1727203825.28150: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13731 1727203825.28240: variable 'ansible_distribution_major_version' from source: facts 13731 1727203825.28243: Evaluated conditional (ansible_distribution_major_version in ['7', '8']): False 13731 1727203825.28247: when evaluation is False, skipping this task 13731 1727203825.28249: _execute() done 13731 1727203825.28253: dumping result to json 13731 1727203825.28257: done dumping result, returning 13731 1727203825.28265: done running TaskExecutor() for managed-node3/TASK: Install yum-utils package [028d2410-947f-82dc-c122-000000000047] 13731 1727203825.28271: sending task result for task 028d2410-947f-82dc-c122-000000000047 skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version in ['7', '8']", "skip_reason": "Conditional result was False" } 13731 1727203825.28401: no more pending results, returning what we have 13731 1727203825.28405: results queue empty 13731 1727203825.28405: checking for any_errors_fatal 13731 1727203825.28410: done checking for any_errors_fatal 13731 1727203825.28411: checking for max_fail_percentage 13731 1727203825.28412: done checking for max_fail_percentage 13731 1727203825.28413: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.28413: done checking to see if all hosts have failed 13731 1727203825.28414: getting the remaining hosts for this loop 13731 1727203825.28416: done getting the remaining hosts for this loop 13731 1727203825.28418: getting the next task for host managed-node3 13731 1727203825.28423: done getting next task for host managed-node3 13731 1727203825.28425: ^ task is: TASK: Enable EPEL 7 13731 1727203825.28428: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.28431: getting variables 13731 1727203825.28432: in VariableManager get_vars() 13731 1727203825.28456: Calling all_inventory to load vars for managed-node3 13731 1727203825.28458: Calling groups_inventory to load vars for managed-node3 13731 1727203825.28461: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.28469: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.28472: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.28474: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.28580: done sending task result for task 028d2410-947f-82dc-c122-000000000047 13731 1727203825.28583: WORKER PROCESS EXITING 13731 1727203825.28593: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.28706: done with get_vars() 13731 1727203825.28714: done getting variables 13731 1727203825.28750: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Enable EPEL 7] *********************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:32 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.016) 0:00:03.518 ***** 13731 1727203825.28768: entering _queue_task() for managed-node3/command 13731 1727203825.28937: worker is 1 (out of 1 available) 13731 1727203825.28949: exiting _queue_task() for managed-node3/command 13731 1727203825.28960: done queuing things up, now waiting for results queue to drain 13731 1727203825.28961: waiting for pending results... 13731 1727203825.29090: running TaskExecutor() for managed-node3/TASK: Enable EPEL 7 13731 1727203825.29150: in run() - task 028d2410-947f-82dc-c122-000000000048 13731 1727203825.29160: variable 'ansible_search_path' from source: unknown 13731 1727203825.29163: variable 'ansible_search_path' from source: unknown 13731 1727203825.29193: calling self._execute() 13731 1727203825.29243: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.29247: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.29254: variable 'omit' from source: magic vars 13731 1727203825.29564: variable 'ansible_distribution' from source: facts 13731 1727203825.29568: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13731 1727203825.29669: variable 'ansible_distribution_major_version' from source: facts 13731 1727203825.29672: Evaluated conditional (ansible_distribution_major_version in ['7', '8']): False 13731 1727203825.29677: when evaluation is False, skipping this task 13731 1727203825.29680: _execute() done 13731 1727203825.29683: dumping result to json 13731 1727203825.29685: done dumping result, returning 13731 1727203825.29688: done running TaskExecutor() for managed-node3/TASK: Enable EPEL 7 [028d2410-947f-82dc-c122-000000000048] 13731 1727203825.29690: sending task result for task 028d2410-947f-82dc-c122-000000000048 13731 1727203825.29795: done sending task result for task 028d2410-947f-82dc-c122-000000000048 13731 1727203825.29800: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version in ['7', '8']", "skip_reason": "Conditional result was False" } 13731 1727203825.29925: no more pending results, returning what we have 13731 1727203825.29928: results queue empty 13731 1727203825.29929: checking for any_errors_fatal 13731 1727203825.29932: done checking for any_errors_fatal 13731 1727203825.29933: checking for max_fail_percentage 13731 1727203825.29934: done checking for max_fail_percentage 13731 1727203825.29935: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.29936: done checking to see if all hosts have failed 13731 1727203825.29936: getting the remaining hosts for this loop 13731 1727203825.29937: done getting the remaining hosts for this loop 13731 1727203825.29940: getting the next task for host managed-node3 13731 1727203825.29945: done getting next task for host managed-node3 13731 1727203825.29947: ^ task is: TASK: Enable EPEL 8 13731 1727203825.29950: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.29953: getting variables 13731 1727203825.29954: in VariableManager get_vars() 13731 1727203825.29979: Calling all_inventory to load vars for managed-node3 13731 1727203825.29984: Calling groups_inventory to load vars for managed-node3 13731 1727203825.29987: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.29995: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.29997: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.29999: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.30196: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.30365: done with get_vars() 13731 1727203825.30371: done getting variables 13731 1727203825.30415: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Enable EPEL 8] *********************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:37 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.016) 0:00:03.535 ***** 13731 1727203825.30441: entering _queue_task() for managed-node3/command 13731 1727203825.30615: worker is 1 (out of 1 available) 13731 1727203825.30626: exiting _queue_task() for managed-node3/command 13731 1727203825.30635: done queuing things up, now waiting for results queue to drain 13731 1727203825.30636: waiting for pending results... 13731 1727203825.30923: running TaskExecutor() for managed-node3/TASK: Enable EPEL 8 13731 1727203825.30927: in run() - task 028d2410-947f-82dc-c122-000000000049 13731 1727203825.30930: variable 'ansible_search_path' from source: unknown 13731 1727203825.30932: variable 'ansible_search_path' from source: unknown 13731 1727203825.30935: calling self._execute() 13731 1727203825.31181: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.31184: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.31187: variable 'omit' from source: magic vars 13731 1727203825.31388: variable 'ansible_distribution' from source: facts 13731 1727203825.31405: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13731 1727203825.31552: variable 'ansible_distribution_major_version' from source: facts 13731 1727203825.31564: Evaluated conditional (ansible_distribution_major_version in ['7', '8']): False 13731 1727203825.31572: when evaluation is False, skipping this task 13731 1727203825.31582: _execute() done 13731 1727203825.31590: dumping result to json 13731 1727203825.31596: done dumping result, returning 13731 1727203825.31605: done running TaskExecutor() for managed-node3/TASK: Enable EPEL 8 [028d2410-947f-82dc-c122-000000000049] 13731 1727203825.31613: sending task result for task 028d2410-947f-82dc-c122-000000000049 13731 1727203825.31740: done sending task result for task 028d2410-947f-82dc-c122-000000000049 13731 1727203825.31743: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version in ['7', '8']", "skip_reason": "Conditional result was False" } 13731 1727203825.31797: no more pending results, returning what we have 13731 1727203825.31804: results queue empty 13731 1727203825.31805: checking for any_errors_fatal 13731 1727203825.31809: done checking for any_errors_fatal 13731 1727203825.31809: checking for max_fail_percentage 13731 1727203825.31811: done checking for max_fail_percentage 13731 1727203825.31811: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.31812: done checking to see if all hosts have failed 13731 1727203825.31813: getting the remaining hosts for this loop 13731 1727203825.31815: done getting the remaining hosts for this loop 13731 1727203825.31818: getting the next task for host managed-node3 13731 1727203825.31825: done getting next task for host managed-node3 13731 1727203825.31827: ^ task is: TASK: Enable EPEL 6 13731 1727203825.31831: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.31834: getting variables 13731 1727203825.31835: in VariableManager get_vars() 13731 1727203825.31865: Calling all_inventory to load vars for managed-node3 13731 1727203825.31868: Calling groups_inventory to load vars for managed-node3 13731 1727203825.31871: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.31881: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.31884: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.31886: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.32018: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.32133: done with get_vars() 13731 1727203825.32139: done getting variables 13731 1727203825.32181: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Enable EPEL 6] *********************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:42 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.017) 0:00:03.553 ***** 13731 1727203825.32201: entering _queue_task() for managed-node3/copy 13731 1727203825.32365: worker is 1 (out of 1 available) 13731 1727203825.32378: exiting _queue_task() for managed-node3/copy 13731 1727203825.32388: done queuing things up, now waiting for results queue to drain 13731 1727203825.32390: waiting for pending results... 13731 1727203825.32524: running TaskExecutor() for managed-node3/TASK: Enable EPEL 6 13731 1727203825.32584: in run() - task 028d2410-947f-82dc-c122-00000000004b 13731 1727203825.32594: variable 'ansible_search_path' from source: unknown 13731 1727203825.32597: variable 'ansible_search_path' from source: unknown 13731 1727203825.32626: calling self._execute() 13731 1727203825.32681: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.32684: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.32693: variable 'omit' from source: magic vars 13731 1727203825.32947: variable 'ansible_distribution' from source: facts 13731 1727203825.32956: Evaluated conditional (ansible_distribution in ['RedHat', 'CentOS']): True 13731 1727203825.33031: variable 'ansible_distribution_major_version' from source: facts 13731 1727203825.33035: Evaluated conditional (ansible_distribution_major_version == '6'): False 13731 1727203825.33038: when evaluation is False, skipping this task 13731 1727203825.33041: _execute() done 13731 1727203825.33043: dumping result to json 13731 1727203825.33047: done dumping result, returning 13731 1727203825.33058: done running TaskExecutor() for managed-node3/TASK: Enable EPEL 6 [028d2410-947f-82dc-c122-00000000004b] 13731 1727203825.33064: sending task result for task 028d2410-947f-82dc-c122-00000000004b 13731 1727203825.33139: done sending task result for task 028d2410-947f-82dc-c122-00000000004b 13731 1727203825.33142: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version == '6'", "skip_reason": "Conditional result was False" } 13731 1727203825.33199: no more pending results, returning what we have 13731 1727203825.33202: results queue empty 13731 1727203825.33203: checking for any_errors_fatal 13731 1727203825.33207: done checking for any_errors_fatal 13731 1727203825.33207: checking for max_fail_percentage 13731 1727203825.33209: done checking for max_fail_percentage 13731 1727203825.33209: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.33210: done checking to see if all hosts have failed 13731 1727203825.33211: getting the remaining hosts for this loop 13731 1727203825.33212: done getting the remaining hosts for this loop 13731 1727203825.33215: getting the next task for host managed-node3 13731 1727203825.33220: done getting next task for host managed-node3 13731 1727203825.33223: ^ task is: TASK: Set network provider to 'nm' 13731 1727203825.33224: ^ state is: HOST STATE: block=2, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.33227: getting variables 13731 1727203825.33228: in VariableManager get_vars() 13731 1727203825.33250: Calling all_inventory to load vars for managed-node3 13731 1727203825.33252: Calling groups_inventory to load vars for managed-node3 13731 1727203825.33255: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.33265: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.33267: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.33270: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.33400: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.33509: done with get_vars() 13731 1727203825.33515: done getting variables 13731 1727203825.33550: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set network provider to 'nm'] ******************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_options_nm.yml:13 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.013) 0:00:03.566 ***** 13731 1727203825.33568: entering _queue_task() for managed-node3/set_fact 13731 1727203825.34006: worker is 1 (out of 1 available) 13731 1727203825.34014: exiting _queue_task() for managed-node3/set_fact 13731 1727203825.34021: done queuing things up, now waiting for results queue to drain 13731 1727203825.34023: waiting for pending results... 13731 1727203825.34148: running TaskExecutor() for managed-node3/TASK: Set network provider to 'nm' 13731 1727203825.34153: in run() - task 028d2410-947f-82dc-c122-000000000007 13731 1727203825.34170: variable 'ansible_search_path' from source: unknown 13731 1727203825.34208: calling self._execute() 13731 1727203825.34288: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.34300: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.34314: variable 'omit' from source: magic vars 13731 1727203825.34440: variable 'omit' from source: magic vars 13731 1727203825.34508: variable 'omit' from source: magic vars 13731 1727203825.34562: variable 'omit' from source: magic vars 13731 1727203825.34618: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203825.34682: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203825.34685: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203825.34706: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203825.34715: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203825.34739: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203825.34742: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.34744: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.34821: Set connection var ansible_pipelining to False 13731 1727203825.34825: Set connection var ansible_shell_type to sh 13731 1727203825.34830: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203825.34835: Set connection var ansible_connection to ssh 13731 1727203825.34840: Set connection var ansible_shell_executable to /bin/sh 13731 1727203825.34845: Set connection var ansible_timeout to 10 13731 1727203825.34865: variable 'ansible_shell_executable' from source: unknown 13731 1727203825.34868: variable 'ansible_connection' from source: unknown 13731 1727203825.34870: variable 'ansible_module_compression' from source: unknown 13731 1727203825.34872: variable 'ansible_shell_type' from source: unknown 13731 1727203825.34880: variable 'ansible_shell_executable' from source: unknown 13731 1727203825.34883: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.34885: variable 'ansible_pipelining' from source: unknown 13731 1727203825.34887: variable 'ansible_timeout' from source: unknown 13731 1727203825.34889: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.35002: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203825.35011: variable 'omit' from source: magic vars 13731 1727203825.35014: starting attempt loop 13731 1727203825.35017: running the handler 13731 1727203825.35028: handler run complete 13731 1727203825.35036: attempt loop complete, returning result 13731 1727203825.35038: _execute() done 13731 1727203825.35041: dumping result to json 13731 1727203825.35043: done dumping result, returning 13731 1727203825.35049: done running TaskExecutor() for managed-node3/TASK: Set network provider to 'nm' [028d2410-947f-82dc-c122-000000000007] 13731 1727203825.35055: sending task result for task 028d2410-947f-82dc-c122-000000000007 13731 1727203825.35132: done sending task result for task 028d2410-947f-82dc-c122-000000000007 13731 1727203825.35134: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "network_provider": "nm" }, "changed": false } 13731 1727203825.35191: no more pending results, returning what we have 13731 1727203825.35194: results queue empty 13731 1727203825.35195: checking for any_errors_fatal 13731 1727203825.35199: done checking for any_errors_fatal 13731 1727203825.35200: checking for max_fail_percentage 13731 1727203825.35201: done checking for max_fail_percentage 13731 1727203825.35202: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.35203: done checking to see if all hosts have failed 13731 1727203825.35203: getting the remaining hosts for this loop 13731 1727203825.35205: done getting the remaining hosts for this loop 13731 1727203825.35208: getting the next task for host managed-node3 13731 1727203825.35212: done getting next task for host managed-node3 13731 1727203825.35214: ^ task is: TASK: meta (flush_handlers) 13731 1727203825.35215: ^ state is: HOST STATE: block=3, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.35219: getting variables 13731 1727203825.35220: in VariableManager get_vars() 13731 1727203825.35242: Calling all_inventory to load vars for managed-node3 13731 1727203825.35246: Calling groups_inventory to load vars for managed-node3 13731 1727203825.35248: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.35256: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.35258: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.35263: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.35372: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.35485: done with get_vars() 13731 1727203825.35492: done getting variables 13731 1727203825.35536: in VariableManager get_vars() 13731 1727203825.35541: Calling all_inventory to load vars for managed-node3 13731 1727203825.35543: Calling groups_inventory to load vars for managed-node3 13731 1727203825.35544: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.35547: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.35548: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.35550: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.35651: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.35762: done with get_vars() 13731 1727203825.35772: done queuing things up, now waiting for results queue to drain 13731 1727203825.35773: results queue empty 13731 1727203825.35774: checking for any_errors_fatal 13731 1727203825.35776: done checking for any_errors_fatal 13731 1727203825.35777: checking for max_fail_percentage 13731 1727203825.35778: done checking for max_fail_percentage 13731 1727203825.35778: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.35778: done checking to see if all hosts have failed 13731 1727203825.35779: getting the remaining hosts for this loop 13731 1727203825.35780: done getting the remaining hosts for this loop 13731 1727203825.35781: getting the next task for host managed-node3 13731 1727203825.35783: done getting next task for host managed-node3 13731 1727203825.35784: ^ task is: TASK: meta (flush_handlers) 13731 1727203825.35785: ^ state is: HOST STATE: block=4, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.35791: getting variables 13731 1727203825.35792: in VariableManager get_vars() 13731 1727203825.35797: Calling all_inventory to load vars for managed-node3 13731 1727203825.35798: Calling groups_inventory to load vars for managed-node3 13731 1727203825.35799: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.35802: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.35804: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.35805: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.35889: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.35993: done with get_vars() 13731 1727203825.35998: done getting variables 13731 1727203825.36025: in VariableManager get_vars() 13731 1727203825.36030: Calling all_inventory to load vars for managed-node3 13731 1727203825.36032: Calling groups_inventory to load vars for managed-node3 13731 1727203825.36033: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.36035: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.36037: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.36038: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.36119: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.36238: done with get_vars() 13731 1727203825.36245: done queuing things up, now waiting for results queue to drain 13731 1727203825.36246: results queue empty 13731 1727203825.36247: checking for any_errors_fatal 13731 1727203825.36247: done checking for any_errors_fatal 13731 1727203825.36248: checking for max_fail_percentage 13731 1727203825.36248: done checking for max_fail_percentage 13731 1727203825.36249: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.36249: done checking to see if all hosts have failed 13731 1727203825.36250: getting the remaining hosts for this loop 13731 1727203825.36250: done getting the remaining hosts for this loop 13731 1727203825.36253: getting the next task for host managed-node3 13731 1727203825.36255: done getting next task for host managed-node3 13731 1727203825.36255: ^ task is: None 13731 1727203825.36256: ^ state is: HOST STATE: block=5, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.36257: done queuing things up, now waiting for results queue to drain 13731 1727203825.36257: results queue empty 13731 1727203825.36257: checking for any_errors_fatal 13731 1727203825.36258: done checking for any_errors_fatal 13731 1727203825.36258: checking for max_fail_percentage 13731 1727203825.36261: done checking for max_fail_percentage 13731 1727203825.36261: checking to see if all hosts have failed and the running result is not ok 13731 1727203825.36261: done checking to see if all hosts have failed 13731 1727203825.36263: getting the next task for host managed-node3 13731 1727203825.36264: done getting next task for host managed-node3 13731 1727203825.36265: ^ task is: None 13731 1727203825.36266: ^ state is: HOST STATE: block=5, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.36302: in VariableManager get_vars() 13731 1727203825.36313: done with get_vars() 13731 1727203825.36317: in VariableManager get_vars() 13731 1727203825.36322: done with get_vars() 13731 1727203825.36325: variable 'omit' from source: magic vars 13731 1727203825.36344: in VariableManager get_vars() 13731 1727203825.36349: done with get_vars() 13731 1727203825.36364: variable 'omit' from source: magic vars PLAY [Play for testing bond options] ******************************************* 13731 1727203825.36516: Loading StrategyModule 'linear' from /usr/local/lib/python3.12/site-packages/ansible/plugins/strategy/linear.py (found_in_cache=True, class_only=False) 13731 1727203825.36539: getting the remaining hosts for this loop 13731 1727203825.36540: done getting the remaining hosts for this loop 13731 1727203825.36542: getting the next task for host managed-node3 13731 1727203825.36543: done getting next task for host managed-node3 13731 1727203825.36545: ^ task is: TASK: Gathering Facts 13731 1727203825.36545: ^ state is: HOST STATE: block=0, task=0, rescue=0, always=0, handlers=0, run_state=0, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=True, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203825.36547: getting variables 13731 1727203825.36547: in VariableManager get_vars() 13731 1727203825.36552: Calling all_inventory to load vars for managed-node3 13731 1727203825.36553: Calling groups_inventory to load vars for managed-node3 13731 1727203825.36555: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203825.36558: Calling all_plugins_play to load vars for managed-node3 13731 1727203825.36568: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203825.36570: Calling groups_plugins_play to load vars for managed-node3 13731 1727203825.36651: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203825.36756: done with get_vars() 13731 1727203825.36763: done getting variables 13731 1727203825.36789: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond_options.yml:3 Tuesday 24 September 2024 14:50:25 -0400 (0:00:00.032) 0:00:03.599 ***** 13731 1727203825.36803: entering _queue_task() for managed-node3/gather_facts 13731 1727203825.36986: worker is 1 (out of 1 available) 13731 1727203825.36997: exiting _queue_task() for managed-node3/gather_facts 13731 1727203825.37007: done queuing things up, now waiting for results queue to drain 13731 1727203825.37008: waiting for pending results... 13731 1727203825.37138: running TaskExecutor() for managed-node3/TASK: Gathering Facts 13731 1727203825.37193: in run() - task 028d2410-947f-82dc-c122-000000000071 13731 1727203825.37205: variable 'ansible_search_path' from source: unknown 13731 1727203825.37234: calling self._execute() 13731 1727203825.37286: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.37289: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.37298: variable 'omit' from source: magic vars 13731 1727203825.37622: variable 'ansible_distribution_major_version' from source: facts 13731 1727203825.37630: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203825.37635: variable 'omit' from source: magic vars 13731 1727203825.37652: variable 'omit' from source: magic vars 13731 1727203825.37678: variable 'omit' from source: magic vars 13731 1727203825.37710: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203825.37734: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203825.37749: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203825.37764: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203825.37773: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203825.37799: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203825.37803: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.37806: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.37869: Set connection var ansible_pipelining to False 13731 1727203825.37873: Set connection var ansible_shell_type to sh 13731 1727203825.37879: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203825.37886: Set connection var ansible_connection to ssh 13731 1727203825.37888: Set connection var ansible_shell_executable to /bin/sh 13731 1727203825.37895: Set connection var ansible_timeout to 10 13731 1727203825.37911: variable 'ansible_shell_executable' from source: unknown 13731 1727203825.37915: variable 'ansible_connection' from source: unknown 13731 1727203825.37919: variable 'ansible_module_compression' from source: unknown 13731 1727203825.37921: variable 'ansible_shell_type' from source: unknown 13731 1727203825.37923: variable 'ansible_shell_executable' from source: unknown 13731 1727203825.37926: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203825.37928: variable 'ansible_pipelining' from source: unknown 13731 1727203825.37930: variable 'ansible_timeout' from source: unknown 13731 1727203825.37932: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203825.38058: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203825.38066: variable 'omit' from source: magic vars 13731 1727203825.38070: starting attempt loop 13731 1727203825.38073: running the handler 13731 1727203825.38088: variable 'ansible_facts' from source: unknown 13731 1727203825.38103: _low_level_execute_command(): starting 13731 1727203825.38124: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203825.38630: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203825.38633: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.38636: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203825.38638: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.38689: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203825.38694: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203825.38702: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203825.38765: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203825.41063: stdout chunk (state=3): >>>/root <<< 13731 1727203825.41200: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203825.41230: stderr chunk (state=3): >>><<< 13731 1727203825.41233: stdout chunk (state=3): >>><<< 13731 1727203825.41252: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203825.41267: _low_level_execute_command(): starting 13731 1727203825.41272: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766 `" && echo ansible-tmp-1727203825.41252-13942-248039094451766="` echo /root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766 `" ) && sleep 0' 13731 1727203825.41715: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203825.41719: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.41743: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.41782: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203825.41794: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203825.41838: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203825.44546: stdout chunk (state=3): >>>ansible-tmp-1727203825.41252-13942-248039094451766=/root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766 <<< 13731 1727203825.44722: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203825.44753: stderr chunk (state=3): >>><<< 13731 1727203825.44756: stdout chunk (state=3): >>><<< 13731 1727203825.44772: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203825.41252-13942-248039094451766=/root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203825.44798: variable 'ansible_module_compression' from source: unknown 13731 1727203825.44841: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.setup-ZIP_DEFLATED 13731 1727203825.44892: variable 'ansible_facts' from source: unknown 13731 1727203825.45024: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/AnsiballZ_setup.py 13731 1727203825.45127: Sending initial data 13731 1727203825.45130: Sent initial data (152 bytes) 13731 1727203825.45590: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203825.45595: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203825.45597: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.45600: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203825.45602: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.45650: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203825.45653: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203825.45698: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203825.47720: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" <<< 13731 1727203825.47724: stderr chunk (state=3): >>>debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203825.47764: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203825.47792: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpj9927xkd /root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/AnsiballZ_setup.py <<< 13731 1727203825.47795: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/AnsiballZ_setup.py" <<< 13731 1727203825.47824: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpj9927xkd" to remote "/root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/AnsiballZ_setup.py" <<< 13731 1727203825.47830: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/AnsiballZ_setup.py" <<< 13731 1727203825.48822: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203825.48867: stderr chunk (state=3): >>><<< 13731 1727203825.48870: stdout chunk (state=3): >>><<< 13731 1727203825.48888: done transferring module to remote 13731 1727203825.48897: _low_level_execute_command(): starting 13731 1727203825.48901: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/ /root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/AnsiballZ_setup.py && sleep 0' 13731 1727203825.49521: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.49528: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203825.49592: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203825.49605: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203825.51454: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203825.51458: stdout chunk (state=3): >>><<< 13731 1727203825.51463: stderr chunk (state=3): >>><<< 13731 1727203825.51563: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203825.51567: _low_level_execute_command(): starting 13731 1727203825.51569: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/AnsiballZ_setup.py && sleep 0' 13731 1727203825.52053: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203825.52068: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203825.52121: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203825.52144: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203825.52258: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203826.31725: stdout chunk (state=3): >>> <<< 13731 1727203826.31813: stdout chunk (state=3): >>>{"ansible_facts": {"ansible_env": {"SHELL": "/bin/bash", "GPG_TTY": "/dev/pts/0", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.12", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.45.138 53526 10.31.47.22 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.45.138 53526 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0"}, "ansible_system": "Linux", "ansible_kernel": "6.11.0-25.el10.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Mon Sep 16 20:35:26 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.12.5", "ansible_fqdn": "managed-node3", "ansible_hostname": "managed-node3", "ansible_nodename": "managed-node3", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "ec25272c602494034078bc876e25857f", "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "Super User", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_is_chroot": false, "ansible_apparmor": {"status": "disabled"}, "ansible_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_hostnqn": "nqn.2014-08.org.nvmexpress:uuid:11e86335-d786-4518-8abc-c9417b351256", "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "10", "ansible_distribution_major_version": "10", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQC83yKiaGAYjfqsqlfpPMYFAYI2IZVpX8WgNfvPCdI/QOKjuzs4b6SbV/Bm0ogtp9Et9STTGoHBvp3tMYQ6i0y/9DHMBxtiHYJ+rIzJ/YNNMqKc+fMqexyOsi7rKKuzfxXrYU8uPBHq5WU3WAZmJxJn+UHZGog8hUnJ8momdJG+aYo9El3Qce4gVdwORcmHZUOa49M8lLCwTovtYArmkGETUVJ+Jk8huVTzYpASWxxcw6zOvUcn52HC6dmNQv/T+k2uW6UW0rybwIrVUlZXRNODrXs8kCGgOx1OI0XYB3FndJOnORF4A9Y6onLo/zUCEaO8Pi19mcfSbo2v+bmotTVk5jcmvR3jhVYJmJE6a+dQpjSZolSqMv8mI9tkztfxM6bJlNSZcTrvZEzu7cbiE38Pp/Ku143n9iGgWNmUQ2FhUDpoWEhoA767VUunE48P8ivpVZ/u5aEyupZSLEuWEvCLmmGPVcf9hVbcXw0n8RFvUwDdD8WARVhN5GJFUVN5JM0=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHm1sMDuVWGACN5pHFDkl6tR93F90YCY4cFGcXcCoQnN+oT963FmBwTMMlfDIm4G2OUATCZuz6QFZP9trAaUzXo=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIAB71QKijTWbanEvrb0ex0kLr0wX6qyv6naldRWNiIFP", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_fibre_channel_wwn": [], "ansible_loadavg": {"1m": 0.65869140625, "5m": 0.37109375, "15m": 0.17529296875}, "ansible_iscsi_iqn": "", "ansible_python": {"version": {"major": 3, "minor": 12, "micro": 5,<<< 13731 1727203826.31913: stdout chunk (state=3): >>> "releaselevel": "final", "serial": 0}, "version_info": [3, 12, 5, "final", 0], "executable": "/usr/bin/python3.12", "has_sslcontext": true, "type": "cpython"}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3531, "ansible_memfree_mb": 2929, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3531, "used": 602, "free": 2929}, "nocache": {"free": 3265, "used": 266}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec25272c-6024-9403-4078-bc876e25857f", "ansible_product_uuid": "ec25272c-6024-9403-4078-bc876e25857f", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "512", "partitions": {"xvda2": {"links": {"ids": [], "uuids": ["973ca870-ed1b-4e56-a8b4-735608119a28"], "labels": [], "masters": []}, "start": "4096", "sectors": "524283871", "sectorsize": 512, "size": "250.00 GB", "uuid": "973ca870-ed1b-4e56-a8b4-735608119a28", "holders": []}, "xvda1": {"links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "start": "2048", "sectors": "2048", "sectorsize": 512, "size": "1.00 MB", "uuid": null, "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda2": ["973ca870-ed1b-4e56-a8b4-735608119a28"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 403, "ansible_lvm": {"lvs": {}, "vgs": {}, "pvs": {}}, "ansible_mounts": [{"mount": "/", "device": "/dev/xvda2", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "dump": 0, "passno": 0, "size_total": 268366229504, "size_available": 261796200448, "block_size": 4096, "block_total": 65519099, "block_available": 63915088, "block_used": 1604011, "inode_total": 131070960, "inode_available": 131027346, "inode_used": 43614, "uuid": "973ca870-ed1b-4e56-a8b4-735608119a28"}], "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "50", "second": "26", "epoch": "1727203826", "epoch_int": "1727203826", "date": "2024-09-24", "time": "14:50:26", "iso8601_micro": "2024-09-24T18:50:26.262327Z", "iso8601": "2024-09-24T18:50:26Z", "iso8601_basic": "20240924T145026262327", "iso8601_basic_short": "20240924T145026", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_local": {}, "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_lsb": {}, "ansible_fips": false, "ansible_interfaces": ["eth0", "lo"], "ansible_eth0": {"device": "eth0", "macaddress": "02:83:38:1a:ae:4d", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.47.22", <<< 13731 1727203826.31949: stdout chunk (state=3): >>>"broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0", "prefix": "22"}, "ipv6": [{"address": "fe80::83:38ff:fe1a:ae4d", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.44.1", "interface": "eth0", "address": "10.31.47.22", "broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0", "prefix": "22", "macaddress": "02:83:38:1a:ae:4d", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.47.22"], "ansible_all_ipv6_addresses": ["fe80::83:38ff:fe1a:ae4d"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.47.22", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::83:38ff:fe1a:ae4d"]}, "ansible_service_mgr": "systemd", "ansible_pkg_mgr": "dnf", "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} <<< 13731 1727203826.34546: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203826.34549: stdout chunk (state=3): >>><<< 13731 1727203826.34552: stderr chunk (state=3): >>><<< 13731 1727203826.34667: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"ansible_env": {"SHELL": "/bin/bash", "GPG_TTY": "/dev/pts/0", "PWD": "/root", "LOGNAME": "root", "XDG_SESSION_TYPE": "tty", "_": "/usr/bin/python3.12", "MOTD_SHOWN": "pam", "HOME": "/root", "LANG": "en_US.UTF-8", "LS_COLORS": "", "SSH_CONNECTION": "10.31.45.138 53526 10.31.47.22 22", "XDG_SESSION_CLASS": "user", "SELINUX_ROLE_REQUESTED": "", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "USER": "root", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "1", "XDG_SESSION_ID": "5", "XDG_RUNTIME_DIR": "/run/user/0", "SSH_CLIENT": "10.31.45.138 53526 22", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "PATH": "/root/.local/bin:/root/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "SELINUX_LEVEL_REQUESTED": "", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "SSH_TTY": "/dev/pts/0"}, "ansible_system": "Linux", "ansible_kernel": "6.11.0-25.el10.x86_64", "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Mon Sep 16 20:35:26 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.12.5", "ansible_fqdn": "managed-node3", "ansible_hostname": "managed-node3", "ansible_nodename": "managed-node3", "ansible_domain": "", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "ec25272c602494034078bc876e25857f", "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "Super User", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_is_chroot": false, "ansible_apparmor": {"status": "disabled"}, "ansible_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": "ttyS0,115200n8"}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,gpt2)/boot/vmlinuz-6.11.0-25.el10.x86_64", "root": "UUID=973ca870-ed1b-4e56-a8b4-735608119a28", "ro": true, "rhgb": true, "crashkernel": "1G-4G:192M,4G-64G:256M,64G-:512M", "net.ifnames": "0", "console": ["tty0", "ttyS0,115200n8"]}, "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_hostnqn": "nqn.2014-08.org.nvmexpress:uuid:11e86335-d786-4518-8abc-c9417b351256", "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "10", "ansible_distribution_major_version": "10", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQC83yKiaGAYjfqsqlfpPMYFAYI2IZVpX8WgNfvPCdI/QOKjuzs4b6SbV/Bm0ogtp9Et9STTGoHBvp3tMYQ6i0y/9DHMBxtiHYJ+rIzJ/YNNMqKc+fMqexyOsi7rKKuzfxXrYU8uPBHq5WU3WAZmJxJn+UHZGog8hUnJ8momdJG+aYo9El3Qce4gVdwORcmHZUOa49M8lLCwTovtYArmkGETUVJ+Jk8huVTzYpASWxxcw6zOvUcn52HC6dmNQv/T+k2uW6UW0rybwIrVUlZXRNODrXs8kCGgOx1OI0XYB3FndJOnORF4A9Y6onLo/zUCEaO8Pi19mcfSbo2v+bmotTVk5jcmvR3jhVYJmJE6a+dQpjSZolSqMv8mI9tkztfxM6bJlNSZcTrvZEzu7cbiE38Pp/Ku143n9iGgWNmUQ2FhUDpoWEhoA767VUunE48P8ivpVZ/u5aEyupZSLEuWEvCLmmGPVcf9hVbcXw0n8RFvUwDdD8WARVhN5GJFUVN5JM0=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHm1sMDuVWGACN5pHFDkl6tR93F90YCY4cFGcXcCoQnN+oT963FmBwTMMlfDIm4G2OUATCZuz6QFZP9trAaUzXo=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIAB71QKijTWbanEvrb0ex0kLr0wX6qyv6naldRWNiIFP", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_fibre_channel_wwn": [], "ansible_loadavg": {"1m": 0.65869140625, "5m": 0.37109375, "15m": 0.17529296875}, "ansible_iscsi_iqn": "", "ansible_python": {"version": {"major": 3, "minor": 12, "micro": 5, "releaselevel": "final", "serial": 0}, "version_info": [3, 12, 5, "final", 0], "executable": "/usr/bin/python3.12", "has_sslcontext": true, "type": "cpython"}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3531, "ansible_memfree_mb": 2929, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3531, "used": 602, "free": 2929}, "nocache": {"free": 3265, "used": 266}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec25272c-6024-9403-4078-bc876e25857f", "ansible_product_uuid": "ec25272c-6024-9403-4078-bc876e25857f", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "512", "partitions": {"xvda2": {"links": {"ids": [], "uuids": ["973ca870-ed1b-4e56-a8b4-735608119a28"], "labels": [], "masters": []}, "start": "4096", "sectors": "524283871", "sectorsize": 512, "size": "250.00 GB", "uuid": "973ca870-ed1b-4e56-a8b4-735608119a28", "holders": []}, "xvda1": {"links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "start": "2048", "sectors": "2048", "sectorsize": 512, "size": "1.00 MB", "uuid": null, "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda2": ["973ca870-ed1b-4e56-a8b4-735608119a28"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 403, "ansible_lvm": {"lvs": {}, "vgs": {}, "pvs": {}}, "ansible_mounts": [{"mount": "/", "device": "/dev/xvda2", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "dump": 0, "passno": 0, "size_total": 268366229504, "size_available": 261796200448, "block_size": 4096, "block_total": 65519099, "block_available": 63915088, "block_used": 1604011, "inode_total": 131070960, "inode_available": 131027346, "inode_used": 43614, "uuid": "973ca870-ed1b-4e56-a8b4-735608119a28"}], "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Tuesday", "weekday_number": "2", "weeknumber": "39", "day": "24", "hour": "14", "minute": "50", "second": "26", "epoch": "1727203826", "epoch_int": "1727203826", "date": "2024-09-24", "time": "14:50:26", "iso8601_micro": "2024-09-24T18:50:26.262327Z", "iso8601": "2024-09-24T18:50:26Z", "iso8601_basic": "20240924T145026262327", "iso8601_basic_short": "20240924T145026", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_local": {}, "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_lsb": {}, "ansible_fips": false, "ansible_interfaces": ["eth0", "lo"], "ansible_eth0": {"device": "eth0", "macaddress": "02:83:38:1a:ae:4d", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.47.22", "broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0", "prefix": "22"}, "ipv6": [{"address": "fe80::83:38ff:fe1a:ae4d", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tls_hw_rx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]", "rx_gro_list": "off", "macsec_hw_offload": "off [fixed]", "rx_udp_gro_forwarding": "off", "hsr_tag_ins_offload": "off [fixed]", "hsr_tag_rm_offload": "off [fixed]", "hsr_fwd_offload": "off [fixed]", "hsr_dup_offload": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.44.1", "interface": "eth0", "address": "10.31.47.22", "broadcast": "10.31.47.255", "netmask": "255.255.252.0", "network": "10.31.44.0", "prefix": "22", "macaddress": "02:83:38:1a:ae:4d", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.47.22"], "ansible_all_ipv6_addresses": ["fe80::83:38ff:fe1a:ae4d"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.47.22", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::83:38ff:fe1a:ae4d"]}, "ansible_service_mgr": "systemd", "ansible_pkg_mgr": "dnf", "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203826.34872: done with _execute_module (ansible.legacy.setup, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.setup', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203826.34961: _low_level_execute_command(): starting 13731 1727203826.34965: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203825.41252-13942-248039094451766/ > /dev/null 2>&1 && sleep 0' 13731 1727203826.35491: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203826.35505: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203826.35518: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203826.35541: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203826.35580: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203826.35599: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203826.35694: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203826.36029: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203826.36104: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203826.38698: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203826.38712: stderr chunk (state=3): >>><<< 13731 1727203826.38731: stdout chunk (state=3): >>><<< 13731 1727203826.38785: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203826.38788: handler run complete 13731 1727203826.38912: variable 'ansible_facts' from source: unknown 13731 1727203826.39086: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.39347: variable 'ansible_facts' from source: unknown 13731 1727203826.39421: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.39552: attempt loop complete, returning result 13731 1727203826.39562: _execute() done 13731 1727203826.39568: dumping result to json 13731 1727203826.39598: done dumping result, returning 13731 1727203826.39609: done running TaskExecutor() for managed-node3/TASK: Gathering Facts [028d2410-947f-82dc-c122-000000000071] 13731 1727203826.39616: sending task result for task 028d2410-947f-82dc-c122-000000000071 13731 1727203826.40302: done sending task result for task 028d2410-947f-82dc-c122-000000000071 13731 1727203826.40306: WORKER PROCESS EXITING ok: [managed-node3] 13731 1727203826.40714: no more pending results, returning what we have 13731 1727203826.40718: results queue empty 13731 1727203826.40719: checking for any_errors_fatal 13731 1727203826.40720: done checking for any_errors_fatal 13731 1727203826.40721: checking for max_fail_percentage 13731 1727203826.40722: done checking for max_fail_percentage 13731 1727203826.40723: checking to see if all hosts have failed and the running result is not ok 13731 1727203826.40724: done checking to see if all hosts have failed 13731 1727203826.40724: getting the remaining hosts for this loop 13731 1727203826.40726: done getting the remaining hosts for this loop 13731 1727203826.40729: getting the next task for host managed-node3 13731 1727203826.40734: done getting next task for host managed-node3 13731 1727203826.40778: ^ task is: TASK: meta (flush_handlers) 13731 1727203826.40781: ^ state is: HOST STATE: block=1, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203826.40785: getting variables 13731 1727203826.40787: in VariableManager get_vars() 13731 1727203826.40807: Calling all_inventory to load vars for managed-node3 13731 1727203826.40809: Calling groups_inventory to load vars for managed-node3 13731 1727203826.40812: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.40821: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.40823: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.40825: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.41217: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.41604: done with get_vars() 13731 1727203826.41614: done getting variables 13731 1727203826.41804: in VariableManager get_vars() 13731 1727203826.41814: Calling all_inventory to load vars for managed-node3 13731 1727203826.41816: Calling groups_inventory to load vars for managed-node3 13731 1727203826.41818: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.41823: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.41826: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.41828: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.42198: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.42633: done with get_vars() 13731 1727203826.42646: done queuing things up, now waiting for results queue to drain 13731 1727203826.42648: results queue empty 13731 1727203826.42649: checking for any_errors_fatal 13731 1727203826.42652: done checking for any_errors_fatal 13731 1727203826.42652: checking for max_fail_percentage 13731 1727203826.42660: done checking for max_fail_percentage 13731 1727203826.42660: checking to see if all hosts have failed and the running result is not ok 13731 1727203826.42661: done checking to see if all hosts have failed 13731 1727203826.42662: getting the remaining hosts for this loop 13731 1727203826.42663: done getting the remaining hosts for this loop 13731 1727203826.42668: getting the next task for host managed-node3 13731 1727203826.42672: done getting next task for host managed-node3 13731 1727203826.42674: ^ task is: TASK: Show playbook name 13731 1727203826.42677: ^ state is: HOST STATE: block=2, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203826.42679: getting variables 13731 1727203826.42680: in VariableManager get_vars() 13731 1727203826.42688: Calling all_inventory to load vars for managed-node3 13731 1727203826.42691: Calling groups_inventory to load vars for managed-node3 13731 1727203826.42693: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.42697: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.42700: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.42703: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.43009: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.43408: done with get_vars() 13731 1727203826.43415: done getting variables 13731 1727203826.43485: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [Show playbook name] ****************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond_options.yml:32 Tuesday 24 September 2024 14:50:26 -0400 (0:00:01.067) 0:00:04.666 ***** 13731 1727203826.43596: entering _queue_task() for managed-node3/debug 13731 1727203826.43598: Creating lock for debug 13731 1727203826.44223: worker is 1 (out of 1 available) 13731 1727203826.44237: exiting _queue_task() for managed-node3/debug 13731 1727203826.44249: done queuing things up, now waiting for results queue to drain 13731 1727203826.44250: waiting for pending results... 13731 1727203826.44694: running TaskExecutor() for managed-node3/TASK: Show playbook name 13731 1727203826.44984: in run() - task 028d2410-947f-82dc-c122-00000000000b 13731 1727203826.44988: variable 'ansible_search_path' from source: unknown 13731 1727203826.44991: calling self._execute() 13731 1727203826.45142: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.45217: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.45245: variable 'omit' from source: magic vars 13731 1727203826.45840: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.45860: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.45872: variable 'omit' from source: magic vars 13731 1727203826.45915: variable 'omit' from source: magic vars 13731 1727203826.45954: variable 'omit' from source: magic vars 13731 1727203826.46000: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203826.46046: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.46071: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203826.46095: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.46112: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.46158: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.46166: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.46174: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.46294: Set connection var ansible_pipelining to False 13731 1727203826.46306: Set connection var ansible_shell_type to sh 13731 1727203826.46315: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.46325: Set connection var ansible_connection to ssh 13731 1727203826.46344: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.46355: Set connection var ansible_timeout to 10 13731 1727203826.46450: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.46452: variable 'ansible_connection' from source: unknown 13731 1727203826.46454: variable 'ansible_module_compression' from source: unknown 13731 1727203826.46456: variable 'ansible_shell_type' from source: unknown 13731 1727203826.46458: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.46459: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.46461: variable 'ansible_pipelining' from source: unknown 13731 1727203826.46463: variable 'ansible_timeout' from source: unknown 13731 1727203826.46464: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.46567: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.46584: variable 'omit' from source: magic vars 13731 1727203826.46593: starting attempt loop 13731 1727203826.46600: running the handler 13731 1727203826.46647: handler run complete 13731 1727203826.46686: attempt loop complete, returning result 13731 1727203826.46693: _execute() done 13731 1727203826.46698: dumping result to json 13731 1727203826.46703: done dumping result, returning 13731 1727203826.46712: done running TaskExecutor() for managed-node3/TASK: Show playbook name [028d2410-947f-82dc-c122-00000000000b] 13731 1727203826.46719: sending task result for task 028d2410-947f-82dc-c122-00000000000b ok: [managed-node3] => {} MSG: this is: playbooks/tests_bond_options.yml 13731 1727203826.46972: no more pending results, returning what we have 13731 1727203826.46978: results queue empty 13731 1727203826.46979: checking for any_errors_fatal 13731 1727203826.46980: done checking for any_errors_fatal 13731 1727203826.46981: checking for max_fail_percentage 13731 1727203826.46983: done checking for max_fail_percentage 13731 1727203826.46983: checking to see if all hosts have failed and the running result is not ok 13731 1727203826.46984: done checking to see if all hosts have failed 13731 1727203826.46985: getting the remaining hosts for this loop 13731 1727203826.46987: done getting the remaining hosts for this loop 13731 1727203826.46990: getting the next task for host managed-node3 13731 1727203826.46998: done getting next task for host managed-node3 13731 1727203826.47001: ^ task is: TASK: Include the task 'run_test.yml' 13731 1727203826.47003: ^ state is: HOST STATE: block=3, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203826.47007: getting variables 13731 1727203826.47009: in VariableManager get_vars() 13731 1727203826.47041: Calling all_inventory to load vars for managed-node3 13731 1727203826.47044: Calling groups_inventory to load vars for managed-node3 13731 1727203826.47048: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.47059: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.47062: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.47066: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.47673: done sending task result for task 028d2410-947f-82dc-c122-00000000000b 13731 1727203826.47680: WORKER PROCESS EXITING 13731 1727203826.47749: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.48139: done with get_vars() 13731 1727203826.48149: done getting variables TASK [Include the task 'run_test.yml'] ***************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond_options.yml:42 Tuesday 24 September 2024 14:50:26 -0400 (0:00:00.048) 0:00:04.715 ***** 13731 1727203826.48432: entering _queue_task() for managed-node3/include_tasks 13731 1727203826.48981: worker is 1 (out of 1 available) 13731 1727203826.48990: exiting _queue_task() for managed-node3/include_tasks 13731 1727203826.49000: done queuing things up, now waiting for results queue to drain 13731 1727203826.49002: waiting for pending results... 13731 1727203826.49338: running TaskExecutor() for managed-node3/TASK: Include the task 'run_test.yml' 13731 1727203826.49447: in run() - task 028d2410-947f-82dc-c122-00000000000d 13731 1727203826.49456: variable 'ansible_search_path' from source: unknown 13731 1727203826.49512: calling self._execute() 13731 1727203826.49665: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.49669: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.49671: variable 'omit' from source: magic vars 13731 1727203826.50007: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.50029: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.50038: _execute() done 13731 1727203826.50044: dumping result to json 13731 1727203826.50050: done dumping result, returning 13731 1727203826.50057: done running TaskExecutor() for managed-node3/TASK: Include the task 'run_test.yml' [028d2410-947f-82dc-c122-00000000000d] 13731 1727203826.50066: sending task result for task 028d2410-947f-82dc-c122-00000000000d 13731 1727203826.50303: no more pending results, returning what we have 13731 1727203826.50310: in VariableManager get_vars() 13731 1727203826.50344: Calling all_inventory to load vars for managed-node3 13731 1727203826.50349: Calling groups_inventory to load vars for managed-node3 13731 1727203826.50353: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.50367: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.50370: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.50373: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.50680: done sending task result for task 028d2410-947f-82dc-c122-00000000000d 13731 1727203826.50683: WORKER PROCESS EXITING 13731 1727203826.50715: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.50909: done with get_vars() 13731 1727203826.50924: variable 'ansible_search_path' from source: unknown 13731 1727203826.50938: we have included files to process 13731 1727203826.50940: generating all_blocks data 13731 1727203826.50941: done generating all_blocks data 13731 1727203826.50942: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml 13731 1727203826.50943: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml 13731 1727203826.50946: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml 13731 1727203826.52256: in VariableManager get_vars() 13731 1727203826.52273: done with get_vars() 13731 1727203826.52686: in VariableManager get_vars() 13731 1727203826.52715: done with get_vars() 13731 1727203826.52755: in VariableManager get_vars() 13731 1727203826.52770: done with get_vars() 13731 1727203826.52940: in VariableManager get_vars() 13731 1727203826.52956: done with get_vars() 13731 1727203826.53010: in VariableManager get_vars() 13731 1727203826.53024: done with get_vars() 13731 1727203826.53924: in VariableManager get_vars() 13731 1727203826.53939: done with get_vars() 13731 1727203826.53950: done processing included file 13731 1727203826.53952: iterating over new_blocks loaded from include file 13731 1727203826.53953: in VariableManager get_vars() 13731 1727203826.53963: done with get_vars() 13731 1727203826.53964: filtering new block on tags 13731 1727203826.54142: done filtering new block on tags 13731 1727203826.54145: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml for managed-node3 13731 1727203826.54150: extending task lists for all hosts with included blocks 13731 1727203826.54186: done extending task lists 13731 1727203826.54187: done processing included files 13731 1727203826.54188: results queue empty 13731 1727203826.54189: checking for any_errors_fatal 13731 1727203826.54193: done checking for any_errors_fatal 13731 1727203826.54193: checking for max_fail_percentage 13731 1727203826.54194: done checking for max_fail_percentage 13731 1727203826.54195: checking to see if all hosts have failed and the running result is not ok 13731 1727203826.54196: done checking to see if all hosts have failed 13731 1727203826.54197: getting the remaining hosts for this loop 13731 1727203826.54198: done getting the remaining hosts for this loop 13731 1727203826.54200: getting the next task for host managed-node3 13731 1727203826.54203: done getting next task for host managed-node3 13731 1727203826.54206: ^ task is: TASK: TEST: {{ lsr_description }} 13731 1727203826.54208: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203826.54210: getting variables 13731 1727203826.54211: in VariableManager get_vars() 13731 1727203826.54219: Calling all_inventory to load vars for managed-node3 13731 1727203826.54221: Calling groups_inventory to load vars for managed-node3 13731 1727203826.54223: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.54343: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.54346: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.54349: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.54627: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.55020: done with get_vars() 13731 1727203826.55030: done getting variables 13731 1727203826.55068: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203826.55360: variable 'lsr_description' from source: include params TASK [TEST: Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:5 Tuesday 24 September 2024 14:50:26 -0400 (0:00:00.070) 0:00:04.786 ***** 13731 1727203826.55505: entering _queue_task() for managed-node3/debug 13731 1727203826.56144: worker is 1 (out of 1 available) 13731 1727203826.56156: exiting _queue_task() for managed-node3/debug 13731 1727203826.56167: done queuing things up, now waiting for results queue to drain 13731 1727203826.56169: waiting for pending results... 13731 1727203826.56289: running TaskExecutor() for managed-node3/TASK: TEST: Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device. 13731 1727203826.56584: in run() - task 028d2410-947f-82dc-c122-000000000088 13731 1727203826.56711: variable 'ansible_search_path' from source: unknown 13731 1727203826.56715: variable 'ansible_search_path' from source: unknown 13731 1727203826.56750: calling self._execute() 13731 1727203826.56990: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.56997: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.57000: variable 'omit' from source: magic vars 13731 1727203826.57868: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.57872: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.57877: variable 'omit' from source: magic vars 13731 1727203826.57879: variable 'omit' from source: magic vars 13731 1727203826.58055: variable 'lsr_description' from source: include params 13731 1727203826.58081: variable 'omit' from source: magic vars 13731 1727203826.58118: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203826.58215: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.58349: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203826.58369: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.58382: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.58412: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.58415: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.58418: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.58631: Set connection var ansible_pipelining to False 13731 1727203826.58636: Set connection var ansible_shell_type to sh 13731 1727203826.58642: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.58648: Set connection var ansible_connection to ssh 13731 1727203826.58654: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.58660: Set connection var ansible_timeout to 10 13731 1727203826.58849: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.58852: variable 'ansible_connection' from source: unknown 13731 1727203826.58855: variable 'ansible_module_compression' from source: unknown 13731 1727203826.58857: variable 'ansible_shell_type' from source: unknown 13731 1727203826.58859: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.58861: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.58863: variable 'ansible_pipelining' from source: unknown 13731 1727203826.58864: variable 'ansible_timeout' from source: unknown 13731 1727203826.58866: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.59067: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.59076: variable 'omit' from source: magic vars 13731 1727203826.59082: starting attempt loop 13731 1727203826.59085: running the handler 13731 1727203826.59242: handler run complete 13731 1727203826.59287: attempt loop complete, returning result 13731 1727203826.59290: _execute() done 13731 1727203826.59292: dumping result to json 13731 1727203826.59294: done dumping result, returning 13731 1727203826.59296: done running TaskExecutor() for managed-node3/TASK: TEST: Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device. [028d2410-947f-82dc-c122-000000000088] 13731 1727203826.59299: sending task result for task 028d2410-947f-82dc-c122-000000000088 ok: [managed-node3] => {} MSG: ########## Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device. ########## 13731 1727203826.59553: no more pending results, returning what we have 13731 1727203826.59557: results queue empty 13731 1727203826.59558: checking for any_errors_fatal 13731 1727203826.59559: done checking for any_errors_fatal 13731 1727203826.59560: checking for max_fail_percentage 13731 1727203826.59561: done checking for max_fail_percentage 13731 1727203826.59562: checking to see if all hosts have failed and the running result is not ok 13731 1727203826.59563: done checking to see if all hosts have failed 13731 1727203826.59563: getting the remaining hosts for this loop 13731 1727203826.59565: done getting the remaining hosts for this loop 13731 1727203826.59568: getting the next task for host managed-node3 13731 1727203826.59574: done getting next task for host managed-node3 13731 1727203826.59578: ^ task is: TASK: Show item 13731 1727203826.59581: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203826.59584: getting variables 13731 1727203826.59585: in VariableManager get_vars() 13731 1727203826.59612: Calling all_inventory to load vars for managed-node3 13731 1727203826.59616: Calling groups_inventory to load vars for managed-node3 13731 1727203826.59619: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.59630: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.59632: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.59634: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.60105: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.60398: done with get_vars() 13731 1727203826.60523: done getting variables 13731 1727203826.60553: done sending task result for task 028d2410-947f-82dc-c122-000000000088 13731 1727203826.60556: WORKER PROCESS EXITING 13731 1727203826.60593: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Show item] *************************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:9 Tuesday 24 September 2024 14:50:26 -0400 (0:00:00.051) 0:00:04.837 ***** 13731 1727203826.60678: entering _queue_task() for managed-node3/debug 13731 1727203826.61289: worker is 1 (out of 1 available) 13731 1727203826.61300: exiting _queue_task() for managed-node3/debug 13731 1727203826.61311: done queuing things up, now waiting for results queue to drain 13731 1727203826.61313: waiting for pending results... 13731 1727203826.62191: running TaskExecutor() for managed-node3/TASK: Show item 13731 1727203826.62196: in run() - task 028d2410-947f-82dc-c122-000000000089 13731 1727203826.62199: variable 'ansible_search_path' from source: unknown 13731 1727203826.62201: variable 'ansible_search_path' from source: unknown 13731 1727203826.62203: variable 'omit' from source: magic vars 13731 1727203826.62343: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.62356: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.62373: variable 'omit' from source: magic vars 13731 1727203826.62994: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.63011: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.63023: variable 'omit' from source: magic vars 13731 1727203826.63063: variable 'omit' from source: magic vars 13731 1727203826.63114: variable 'item' from source: unknown 13731 1727203826.63191: variable 'item' from source: unknown 13731 1727203826.63213: variable 'omit' from source: magic vars 13731 1727203826.63255: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203826.63296: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.63317: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203826.63335: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.63348: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.63383: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.63391: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.63398: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.63981: Set connection var ansible_pipelining to False 13731 1727203826.63984: Set connection var ansible_shell_type to sh 13731 1727203826.63987: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.63989: Set connection var ansible_connection to ssh 13731 1727203826.63990: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.63992: Set connection var ansible_timeout to 10 13731 1727203826.63994: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.63996: variable 'ansible_connection' from source: unknown 13731 1727203826.63998: variable 'ansible_module_compression' from source: unknown 13731 1727203826.64000: variable 'ansible_shell_type' from source: unknown 13731 1727203826.64001: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.64003: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.64005: variable 'ansible_pipelining' from source: unknown 13731 1727203826.64007: variable 'ansible_timeout' from source: unknown 13731 1727203826.64009: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.64040: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.64480: variable 'omit' from source: magic vars 13731 1727203826.64483: starting attempt loop 13731 1727203826.64486: running the handler 13731 1727203826.64488: variable 'lsr_description' from source: include params 13731 1727203826.64490: variable 'lsr_description' from source: include params 13731 1727203826.64492: handler run complete 13731 1727203826.64494: attempt loop complete, returning result 13731 1727203826.64496: variable 'item' from source: unknown 13731 1727203826.64578: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_description) => { "ansible_loop_var": "item", "item": "lsr_description", "lsr_description": "Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device." } 13731 1727203826.65081: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.65084: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.65087: variable 'omit' from source: magic vars 13731 1727203826.65366: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.65428: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.65437: variable 'omit' from source: magic vars 13731 1727203826.65504: variable 'omit' from source: magic vars 13731 1727203826.65573: variable 'item' from source: unknown 13731 1727203826.65846: variable 'item' from source: unknown 13731 1727203826.66180: variable 'omit' from source: magic vars 13731 1727203826.66183: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.66185: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.66188: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.66190: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.66192: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.66194: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.66195: Set connection var ansible_pipelining to False 13731 1727203826.66197: Set connection var ansible_shell_type to sh 13731 1727203826.66199: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.66200: Set connection var ansible_connection to ssh 13731 1727203826.66202: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.66204: Set connection var ansible_timeout to 10 13731 1727203826.66205: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.66207: variable 'ansible_connection' from source: unknown 13731 1727203826.66209: variable 'ansible_module_compression' from source: unknown 13731 1727203826.66211: variable 'ansible_shell_type' from source: unknown 13731 1727203826.66212: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.66214: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.66484: variable 'ansible_pipelining' from source: unknown 13731 1727203826.66494: variable 'ansible_timeout' from source: unknown 13731 1727203826.66502: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.66885: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.66941: variable 'omit' from source: magic vars 13731 1727203826.66949: starting attempt loop 13731 1727203826.66957: running the handler 13731 1727203826.66991: variable 'lsr_setup' from source: include params 13731 1727203826.67126: variable 'lsr_setup' from source: include params 13731 1727203826.67180: handler run complete 13731 1727203826.67226: attempt loop complete, returning result 13731 1727203826.67240: variable 'item' from source: unknown 13731 1727203826.67311: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_setup) => { "ansible_loop_var": "item", "item": "lsr_setup", "lsr_setup": [ "tasks/create_test_interfaces_with_dhcp.yml", "tasks/assert_dhcp_device_present.yml" ] } 13731 1727203826.67629: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.67633: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.67635: variable 'omit' from source: magic vars 13731 1727203826.67769: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.67783: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.67792: variable 'omit' from source: magic vars 13731 1727203826.67810: variable 'omit' from source: magic vars 13731 1727203826.67861: variable 'item' from source: unknown 13731 1727203826.67929: variable 'item' from source: unknown 13731 1727203826.67979: variable 'omit' from source: magic vars 13731 1727203826.67983: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.67985: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.67996: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.68012: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.68019: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.68063: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.68384: Set connection var ansible_pipelining to False 13731 1727203826.68387: Set connection var ansible_shell_type to sh 13731 1727203826.68389: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.68391: Set connection var ansible_connection to ssh 13731 1727203826.68393: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.68394: Set connection var ansible_timeout to 10 13731 1727203826.68396: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.68397: variable 'ansible_connection' from source: unknown 13731 1727203826.68399: variable 'ansible_module_compression' from source: unknown 13731 1727203826.68400: variable 'ansible_shell_type' from source: unknown 13731 1727203826.68402: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.68404: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.68405: variable 'ansible_pipelining' from source: unknown 13731 1727203826.68407: variable 'ansible_timeout' from source: unknown 13731 1727203826.68408: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.68555: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.68569: variable 'omit' from source: magic vars 13731 1727203826.68587: starting attempt loop 13731 1727203826.68782: running the handler 13731 1727203826.68785: variable 'lsr_test' from source: include params 13731 1727203826.68788: variable 'lsr_test' from source: include params 13731 1727203826.68954: handler run complete 13731 1727203826.69148: attempt loop complete, returning result 13731 1727203826.69151: variable 'item' from source: unknown 13731 1727203826.69482: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_test) => { "ansible_loop_var": "item", "item": "lsr_test", "lsr_test": [ "tasks/create_bond_profile.yml" ] } 13731 1727203826.69551: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.69554: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.70481: variable 'omit' from source: magic vars 13731 1727203826.70486: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.70489: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.70491: variable 'omit' from source: magic vars 13731 1727203826.70493: variable 'omit' from source: magic vars 13731 1727203826.71058: variable 'item' from source: unknown 13731 1727203826.71061: variable 'item' from source: unknown 13731 1727203826.71481: variable 'omit' from source: magic vars 13731 1727203826.71487: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.71489: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.71492: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.71494: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.71496: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.71498: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.71723: Set connection var ansible_pipelining to False 13731 1727203826.71726: Set connection var ansible_shell_type to sh 13731 1727203826.71728: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.71731: Set connection var ansible_connection to ssh 13731 1727203826.71733: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.71735: Set connection var ansible_timeout to 10 13731 1727203826.71736: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.71738: variable 'ansible_connection' from source: unknown 13731 1727203826.71740: variable 'ansible_module_compression' from source: unknown 13731 1727203826.71742: variable 'ansible_shell_type' from source: unknown 13731 1727203826.71744: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.71746: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.71748: variable 'ansible_pipelining' from source: unknown 13731 1727203826.71749: variable 'ansible_timeout' from source: unknown 13731 1727203826.71751: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.71981: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.72062: variable 'omit' from source: magic vars 13731 1727203826.72072: starting attempt loop 13731 1727203826.72082: running the handler 13731 1727203826.72106: variable 'lsr_assert' from source: include params 13731 1727203826.72303: variable 'lsr_assert' from source: include params 13731 1727203826.72306: handler run complete 13731 1727203826.72383: attempt loop complete, returning result 13731 1727203826.72403: variable 'item' from source: unknown 13731 1727203826.72552: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_assert) => { "ansible_loop_var": "item", "item": "lsr_assert", "lsr_assert": [ "tasks/assert_controller_device_present.yml", "tasks/assert_bond_port_profile_present.yml", "tasks/assert_bond_options.yml" ] } 13731 1727203826.72851: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.73085: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.73092: variable 'omit' from source: magic vars 13731 1727203826.73166: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.73480: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.73484: variable 'omit' from source: magic vars 13731 1727203826.73486: variable 'omit' from source: magic vars 13731 1727203826.73488: variable 'item' from source: unknown 13731 1727203826.73525: variable 'item' from source: unknown 13731 1727203826.73544: variable 'omit' from source: magic vars 13731 1727203826.73632: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.73645: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.73656: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.73673: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.73684: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.73692: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.73936: Set connection var ansible_pipelining to False 13731 1727203826.73939: Set connection var ansible_shell_type to sh 13731 1727203826.73941: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.73944: Set connection var ansible_connection to ssh 13731 1727203826.73946: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.73948: Set connection var ansible_timeout to 10 13731 1727203826.73986: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.74063: variable 'ansible_connection' from source: unknown 13731 1727203826.74066: variable 'ansible_module_compression' from source: unknown 13731 1727203826.74068: variable 'ansible_shell_type' from source: unknown 13731 1727203826.74154: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.74157: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.74159: variable 'ansible_pipelining' from source: unknown 13731 1727203826.74161: variable 'ansible_timeout' from source: unknown 13731 1727203826.74163: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.74301: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.74315: variable 'omit' from source: magic vars 13731 1727203826.74482: starting attempt loop 13731 1727203826.74486: running the handler 13731 1727203826.74612: handler run complete 13731 1727203826.74629: attempt loop complete, returning result 13731 1727203826.74648: variable 'item' from source: unknown 13731 1727203826.74829: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_assert_when) => { "ansible_loop_var": "item", "item": "lsr_assert_when", "lsr_assert_when": "VARIABLE IS NOT DEFINED!: 'lsr_assert_when' is undefined" } 13731 1727203826.76067: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.76070: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.76073: variable 'omit' from source: magic vars 13731 1727203826.76076: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.76284: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.76287: variable 'omit' from source: magic vars 13731 1727203826.76289: variable 'omit' from source: magic vars 13731 1727203826.76291: variable 'item' from source: unknown 13731 1727203826.76733: variable 'item' from source: unknown 13731 1727203826.76736: variable 'omit' from source: magic vars 13731 1727203826.76739: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.76741: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.76743: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.76745: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.76746: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.76748: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.77060: Set connection var ansible_pipelining to False 13731 1727203826.77066: Set connection var ansible_shell_type to sh 13731 1727203826.77068: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.77071: Set connection var ansible_connection to ssh 13731 1727203826.77073: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.77077: Set connection var ansible_timeout to 10 13731 1727203826.77079: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.77081: variable 'ansible_connection' from source: unknown 13731 1727203826.77083: variable 'ansible_module_compression' from source: unknown 13731 1727203826.77085: variable 'ansible_shell_type' from source: unknown 13731 1727203826.77086: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.77088: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.77090: variable 'ansible_pipelining' from source: unknown 13731 1727203826.77091: variable 'ansible_timeout' from source: unknown 13731 1727203826.77093: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.77249: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.77257: variable 'omit' from source: magic vars 13731 1727203826.77260: starting attempt loop 13731 1727203826.77265: running the handler 13731 1727203826.77285: variable 'lsr_fail_debug' from source: play vars 13731 1727203826.77353: variable 'lsr_fail_debug' from source: play vars 13731 1727203826.77389: handler run complete 13731 1727203826.77392: attempt loop complete, returning result 13731 1727203826.77405: variable 'item' from source: unknown 13731 1727203826.77581: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_fail_debug) => { "ansible_loop_var": "item", "item": "lsr_fail_debug", "lsr_fail_debug": [ "__network_connections_result" ] } 13731 1727203826.77655: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.77658: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.77660: variable 'omit' from source: magic vars 13731 1727203826.77884: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.77887: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.77890: variable 'omit' from source: magic vars 13731 1727203826.77892: variable 'omit' from source: magic vars 13731 1727203826.77894: variable 'item' from source: unknown 13731 1727203826.77896: variable 'item' from source: unknown 13731 1727203826.77898: variable 'omit' from source: magic vars 13731 1727203826.77899: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.77901: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.77907: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.77910: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.77912: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.77914: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.78056: Set connection var ansible_pipelining to False 13731 1727203826.78059: Set connection var ansible_shell_type to sh 13731 1727203826.78065: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.78067: Set connection var ansible_connection to ssh 13731 1727203826.78070: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.78072: Set connection var ansible_timeout to 10 13731 1727203826.78074: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.78079: variable 'ansible_connection' from source: unknown 13731 1727203826.78081: variable 'ansible_module_compression' from source: unknown 13731 1727203826.78083: variable 'ansible_shell_type' from source: unknown 13731 1727203826.78085: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.78086: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.78088: variable 'ansible_pipelining' from source: unknown 13731 1727203826.78090: variable 'ansible_timeout' from source: unknown 13731 1727203826.78092: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.78302: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.78309: variable 'omit' from source: magic vars 13731 1727203826.78314: starting attempt loop 13731 1727203826.78481: running the handler 13731 1727203826.78484: variable 'lsr_cleanup' from source: include params 13731 1727203826.78487: variable 'lsr_cleanup' from source: include params 13731 1727203826.78489: handler run complete 13731 1727203826.78491: attempt loop complete, returning result 13731 1727203826.78493: variable 'item' from source: unknown 13731 1727203826.78536: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_cleanup) => { "ansible_loop_var": "item", "item": "lsr_cleanup", "lsr_cleanup": [ "tasks/cleanup_bond_profile+device.yml", "tasks/remove_test_interfaces_with_dhcp.yml" ] } 13731 1727203826.78742: dumping result to json 13731 1727203826.78745: done dumping result, returning 13731 1727203826.78835: done running TaskExecutor() for managed-node3/TASK: Show item [028d2410-947f-82dc-c122-000000000089] 13731 1727203826.78841: sending task result for task 028d2410-947f-82dc-c122-000000000089 13731 1727203826.78912: done sending task result for task 028d2410-947f-82dc-c122-000000000089 13731 1727203826.78915: WORKER PROCESS EXITING 13731 1727203826.79005: no more pending results, returning what we have 13731 1727203826.79009: results queue empty 13731 1727203826.79010: checking for any_errors_fatal 13731 1727203826.79015: done checking for any_errors_fatal 13731 1727203826.79016: checking for max_fail_percentage 13731 1727203826.79017: done checking for max_fail_percentage 13731 1727203826.79018: checking to see if all hosts have failed and the running result is not ok 13731 1727203826.79018: done checking to see if all hosts have failed 13731 1727203826.79019: getting the remaining hosts for this loop 13731 1727203826.79020: done getting the remaining hosts for this loop 13731 1727203826.79023: getting the next task for host managed-node3 13731 1727203826.79029: done getting next task for host managed-node3 13731 1727203826.79031: ^ task is: TASK: Include the task 'show_interfaces.yml' 13731 1727203826.79034: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203826.79038: getting variables 13731 1727203826.79040: in VariableManager get_vars() 13731 1727203826.79066: Calling all_inventory to load vars for managed-node3 13731 1727203826.79068: Calling groups_inventory to load vars for managed-node3 13731 1727203826.79072: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.79083: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.79086: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.79088: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.79346: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.79690: done with get_vars() 13731 1727203826.79702: done getting variables TASK [Include the task 'show_interfaces.yml'] ********************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:21 Tuesday 24 September 2024 14:50:26 -0400 (0:00:00.192) 0:00:05.030 ***** 13731 1727203826.79912: entering _queue_task() for managed-node3/include_tasks 13731 1727203826.80637: worker is 1 (out of 1 available) 13731 1727203826.80650: exiting _queue_task() for managed-node3/include_tasks 13731 1727203826.80661: done queuing things up, now waiting for results queue to drain 13731 1727203826.80663: waiting for pending results... 13731 1727203826.81196: running TaskExecutor() for managed-node3/TASK: Include the task 'show_interfaces.yml' 13731 1727203826.81401: in run() - task 028d2410-947f-82dc-c122-00000000008a 13731 1727203826.81425: variable 'ansible_search_path' from source: unknown 13731 1727203826.81508: variable 'ansible_search_path' from source: unknown 13731 1727203826.81617: calling self._execute() 13731 1727203826.81691: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.81810: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.81945: variable 'omit' from source: magic vars 13731 1727203826.82567: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.82589: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.82637: _execute() done 13731 1727203826.82648: dumping result to json 13731 1727203826.82656: done dumping result, returning 13731 1727203826.82667: done running TaskExecutor() for managed-node3/TASK: Include the task 'show_interfaces.yml' [028d2410-947f-82dc-c122-00000000008a] 13731 1727203826.82680: sending task result for task 028d2410-947f-82dc-c122-00000000008a 13731 1727203826.82881: done sending task result for task 028d2410-947f-82dc-c122-00000000008a 13731 1727203826.82884: WORKER PROCESS EXITING 13731 1727203826.82913: no more pending results, returning what we have 13731 1727203826.82919: in VariableManager get_vars() 13731 1727203826.82958: Calling all_inventory to load vars for managed-node3 13731 1727203826.82961: Calling groups_inventory to load vars for managed-node3 13731 1727203826.82965: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.82981: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.82985: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.82988: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.83387: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.83573: done with get_vars() 13731 1727203826.83583: variable 'ansible_search_path' from source: unknown 13731 1727203826.83584: variable 'ansible_search_path' from source: unknown 13731 1727203826.83625: we have included files to process 13731 1727203826.83626: generating all_blocks data 13731 1727203826.83628: done generating all_blocks data 13731 1727203826.83632: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml 13731 1727203826.83633: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml 13731 1727203826.83635: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml 13731 1727203826.83780: in VariableManager get_vars() 13731 1727203826.83795: done with get_vars() 13731 1727203826.83885: done processing included file 13731 1727203826.83887: iterating over new_blocks loaded from include file 13731 1727203826.83888: in VariableManager get_vars() 13731 1727203826.83899: done with get_vars() 13731 1727203826.83900: filtering new block on tags 13731 1727203826.83926: done filtering new block on tags 13731 1727203826.83928: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml for managed-node3 13731 1727203826.83933: extending task lists for all hosts with included blocks 13731 1727203826.84335: done extending task lists 13731 1727203826.84336: done processing included files 13731 1727203826.84337: results queue empty 13731 1727203826.84338: checking for any_errors_fatal 13731 1727203826.84343: done checking for any_errors_fatal 13731 1727203826.84344: checking for max_fail_percentage 13731 1727203826.84345: done checking for max_fail_percentage 13731 1727203826.84346: checking to see if all hosts have failed and the running result is not ok 13731 1727203826.84347: done checking to see if all hosts have failed 13731 1727203826.84348: getting the remaining hosts for this loop 13731 1727203826.84349: done getting the remaining hosts for this loop 13731 1727203826.84351: getting the next task for host managed-node3 13731 1727203826.84354: done getting next task for host managed-node3 13731 1727203826.84357: ^ task is: TASK: Include the task 'get_current_interfaces.yml' 13731 1727203826.84359: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203826.84361: getting variables 13731 1727203826.84362: in VariableManager get_vars() 13731 1727203826.84370: Calling all_inventory to load vars for managed-node3 13731 1727203826.84372: Calling groups_inventory to load vars for managed-node3 13731 1727203826.84374: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.84381: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.84383: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.84386: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.84723: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.85029: done with get_vars() 13731 1727203826.85038: done getting variables TASK [Include the task 'get_current_interfaces.yml'] *************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml:3 Tuesday 24 September 2024 14:50:26 -0400 (0:00:00.051) 0:00:05.082 ***** 13731 1727203826.85112: entering _queue_task() for managed-node3/include_tasks 13731 1727203826.85381: worker is 1 (out of 1 available) 13731 1727203826.85393: exiting _queue_task() for managed-node3/include_tasks 13731 1727203826.85404: done queuing things up, now waiting for results queue to drain 13731 1727203826.85406: waiting for pending results... 13731 1727203826.85794: running TaskExecutor() for managed-node3/TASK: Include the task 'get_current_interfaces.yml' 13731 1727203826.85800: in run() - task 028d2410-947f-82dc-c122-0000000000b1 13731 1727203826.85802: variable 'ansible_search_path' from source: unknown 13731 1727203826.85804: variable 'ansible_search_path' from source: unknown 13731 1727203826.85817: calling self._execute() 13731 1727203826.85913: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.85924: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.85939: variable 'omit' from source: magic vars 13731 1727203826.86444: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.86460: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.86474: _execute() done 13731 1727203826.86484: dumping result to json 13731 1727203826.86491: done dumping result, returning 13731 1727203826.86500: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_current_interfaces.yml' [028d2410-947f-82dc-c122-0000000000b1] 13731 1727203826.86509: sending task result for task 028d2410-947f-82dc-c122-0000000000b1 13731 1727203826.86626: no more pending results, returning what we have 13731 1727203826.86632: in VariableManager get_vars() 13731 1727203826.86666: Calling all_inventory to load vars for managed-node3 13731 1727203826.86670: Calling groups_inventory to load vars for managed-node3 13731 1727203826.86673: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.86690: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.86693: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.86696: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.87067: done sending task result for task 028d2410-947f-82dc-c122-0000000000b1 13731 1727203826.87071: WORKER PROCESS EXITING 13731 1727203826.87095: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.87532: done with get_vars() 13731 1727203826.87540: variable 'ansible_search_path' from source: unknown 13731 1727203826.87541: variable 'ansible_search_path' from source: unknown 13731 1727203826.87577: we have included files to process 13731 1727203826.87578: generating all_blocks data 13731 1727203826.87580: done generating all_blocks data 13731 1727203826.87582: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml 13731 1727203826.87583: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml 13731 1727203826.87585: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml 13731 1727203826.88089: done processing included file 13731 1727203826.88091: iterating over new_blocks loaded from include file 13731 1727203826.88093: in VariableManager get_vars() 13731 1727203826.88107: done with get_vars() 13731 1727203826.88109: filtering new block on tags 13731 1727203826.88143: done filtering new block on tags 13731 1727203826.88146: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml for managed-node3 13731 1727203826.88151: extending task lists for all hosts with included blocks 13731 1727203826.88520: done extending task lists 13731 1727203826.88522: done processing included files 13731 1727203826.88523: results queue empty 13731 1727203826.88523: checking for any_errors_fatal 13731 1727203826.88527: done checking for any_errors_fatal 13731 1727203826.88527: checking for max_fail_percentage 13731 1727203826.88528: done checking for max_fail_percentage 13731 1727203826.88529: checking to see if all hosts have failed and the running result is not ok 13731 1727203826.88530: done checking to see if all hosts have failed 13731 1727203826.88531: getting the remaining hosts for this loop 13731 1727203826.88532: done getting the remaining hosts for this loop 13731 1727203826.88535: getting the next task for host managed-node3 13731 1727203826.88539: done getting next task for host managed-node3 13731 1727203826.88541: ^ task is: TASK: Gather current interface info 13731 1727203826.88544: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203826.88546: getting variables 13731 1727203826.88547: in VariableManager get_vars() 13731 1727203826.88556: Calling all_inventory to load vars for managed-node3 13731 1727203826.88559: Calling groups_inventory to load vars for managed-node3 13731 1727203826.88561: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203826.88569: Calling all_plugins_play to load vars for managed-node3 13731 1727203826.88572: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203826.88575: Calling groups_plugins_play to load vars for managed-node3 13731 1727203826.89597: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203826.90183: done with get_vars() 13731 1727203826.90194: done getting variables 13731 1727203826.90232: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Gather current interface info] ******************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml:3 Tuesday 24 September 2024 14:50:26 -0400 (0:00:00.051) 0:00:05.133 ***** 13731 1727203826.90261: entering _queue_task() for managed-node3/command 13731 1727203826.90956: worker is 1 (out of 1 available) 13731 1727203826.90971: exiting _queue_task() for managed-node3/command 13731 1727203826.91390: done queuing things up, now waiting for results queue to drain 13731 1727203826.91392: waiting for pending results... 13731 1727203826.91415: running TaskExecutor() for managed-node3/TASK: Gather current interface info 13731 1727203826.91643: in run() - task 028d2410-947f-82dc-c122-0000000000ec 13731 1727203826.91666: variable 'ansible_search_path' from source: unknown 13731 1727203826.91711: variable 'ansible_search_path' from source: unknown 13731 1727203826.91751: calling self._execute() 13731 1727203826.91897: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.92040: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.92055: variable 'omit' from source: magic vars 13731 1727203826.92749: variable 'ansible_distribution_major_version' from source: facts 13731 1727203826.93009: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203826.93013: variable 'omit' from source: magic vars 13731 1727203826.93015: variable 'omit' from source: magic vars 13731 1727203826.93017: variable 'omit' from source: magic vars 13731 1727203826.93128: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203826.93172: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203826.93247: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203826.93352: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.93373: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203826.93409: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203826.93419: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.93446: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.93783: Set connection var ansible_pipelining to False 13731 1727203826.93787: Set connection var ansible_shell_type to sh 13731 1727203826.93789: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203826.93791: Set connection var ansible_connection to ssh 13731 1727203826.93793: Set connection var ansible_shell_executable to /bin/sh 13731 1727203826.93795: Set connection var ansible_timeout to 10 13731 1727203826.93797: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.93798: variable 'ansible_connection' from source: unknown 13731 1727203826.93801: variable 'ansible_module_compression' from source: unknown 13731 1727203826.93803: variable 'ansible_shell_type' from source: unknown 13731 1727203826.93805: variable 'ansible_shell_executable' from source: unknown 13731 1727203826.93807: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203826.93809: variable 'ansible_pipelining' from source: unknown 13731 1727203826.93810: variable 'ansible_timeout' from source: unknown 13731 1727203826.93812: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203826.94061: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203826.94229: variable 'omit' from source: magic vars 13731 1727203826.94241: starting attempt loop 13731 1727203826.94248: running the handler 13731 1727203826.94273: _low_level_execute_command(): starting 13731 1727203826.94289: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203826.95639: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203826.95657: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203826.95846: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203826.95965: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203826.97797: stdout chunk (state=3): >>>/root <<< 13731 1727203826.97833: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203826.97836: stdout chunk (state=3): >>><<< 13731 1727203826.97839: stderr chunk (state=3): >>><<< 13731 1727203826.97861: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203826.97932: _low_level_execute_command(): starting 13731 1727203826.97991: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971 `" && echo ansible-tmp-1727203826.9791784-14079-25446266985971="` echo /root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971 `" ) && sleep 0' 13731 1727203826.99318: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203826.99322: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203826.99394: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203826.99409: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203826.99435: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203826.99472: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203826.99545: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203826.99693: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203826.99743: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.01583: stdout chunk (state=3): >>>ansible-tmp-1727203826.9791784-14079-25446266985971=/root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971 <<< 13731 1727203827.01723: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203827.01735: stdout chunk (state=3): >>><<< 13731 1727203827.01755: stderr chunk (state=3): >>><<< 13731 1727203827.01773: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203826.9791784-14079-25446266985971=/root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203827.01862: variable 'ansible_module_compression' from source: unknown 13731 1727203827.02068: ANSIBALLZ: Using generic lock for ansible.legacy.command 13731 1727203827.02071: ANSIBALLZ: Acquiring lock 13731 1727203827.02074: ANSIBALLZ: Lock acquired: 140078454804688 13731 1727203827.02077: ANSIBALLZ: Creating module 13731 1727203827.22981: ANSIBALLZ: Writing module into payload 13731 1727203827.23126: ANSIBALLZ: Writing module 13731 1727203827.23157: ANSIBALLZ: Renaming module 13731 1727203827.23181: ANSIBALLZ: Done creating module 13731 1727203827.23204: variable 'ansible_facts' from source: unknown 13731 1727203827.23586: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/AnsiballZ_command.py 13731 1727203827.23718: Sending initial data 13731 1727203827.23726: Sent initial data (155 bytes) 13731 1727203827.24938: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203827.24954: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203827.25011: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.25118: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203827.25342: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.25356: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.26993: stderr chunk (state=3): >>>debug2: Remote version: 3 <<< 13731 1727203827.27009: stderr chunk (state=3): >>>debug2: Server supports extension "posix-rename@openssh.com" revision 1 <<< 13731 1727203827.27023: stderr chunk (state=3): >>>debug2: Server supports extension "statvfs@openssh.com" revision 2 <<< 13731 1727203827.27046: stderr chunk (state=3): >>>debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203827.27080: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203827.27142: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpe4555w72 /root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/AnsiballZ_command.py <<< 13731 1727203827.27153: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/AnsiballZ_command.py" <<< 13731 1727203827.27194: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpe4555w72" to remote "/root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/AnsiballZ_command.py" <<< 13731 1727203827.27952: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203827.28016: stderr chunk (state=3): >>><<< 13731 1727203827.28081: stdout chunk (state=3): >>><<< 13731 1727203827.28084: done transferring module to remote 13731 1727203827.28087: _low_level_execute_command(): starting 13731 1727203827.28090: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/ /root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/AnsiballZ_command.py && sleep 0' 13731 1727203827.28925: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203827.28981: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.29063: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203827.29097: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203827.29131: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.29291: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.31039: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203827.31043: stdout chunk (state=3): >>><<< 13731 1727203827.31045: stderr chunk (state=3): >>><<< 13731 1727203827.31061: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203827.31142: _low_level_execute_command(): starting 13731 1727203827.31145: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/AnsiballZ_command.py && sleep 0' 13731 1727203827.31661: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203827.31677: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203827.31693: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203827.31708: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203827.31797: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.31818: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203827.31834: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203827.31857: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.31930: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.47437: stdout chunk (state=3): >>> {"changed": true, "stdout": "bonding_masters\neth0\nlo", "stderr": "", "rc": 0, "cmd": ["ls", "-1"], "start": "2024-09-24 14:50:27.470268", "end": "2024-09-24 14:50:27.473623", "delta": "0:00:00.003355", "msg": "", "invocation": {"module_args": {"chdir": "/sys/class/net", "_raw_params": "ls -1", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203827.48983: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203827.49008: stderr chunk (state=3): >>><<< 13731 1727203827.49012: stdout chunk (state=3): >>><<< 13731 1727203827.49031: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "bonding_masters\neth0\nlo", "stderr": "", "rc": 0, "cmd": ["ls", "-1"], "start": "2024-09-24 14:50:27.470268", "end": "2024-09-24 14:50:27.473623", "delta": "0:00:00.003355", "msg": "", "invocation": {"module_args": {"chdir": "/sys/class/net", "_raw_params": "ls -1", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203827.49061: done with _execute_module (ansible.legacy.command, {'chdir': '/sys/class/net', '_raw_params': 'ls -1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203827.49070: _low_level_execute_command(): starting 13731 1727203827.49077: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203826.9791784-14079-25446266985971/ > /dev/null 2>&1 && sleep 0' 13731 1727203827.49551: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203827.49555: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.49557: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203827.49564: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.49653: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.49683: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.51511: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203827.51534: stderr chunk (state=3): >>><<< 13731 1727203827.51537: stdout chunk (state=3): >>><<< 13731 1727203827.51552: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203827.51557: handler run complete 13731 1727203827.51583: Evaluated conditional (False): False 13731 1727203827.51592: attempt loop complete, returning result 13731 1727203827.51594: _execute() done 13731 1727203827.51597: dumping result to json 13731 1727203827.51601: done dumping result, returning 13731 1727203827.51609: done running TaskExecutor() for managed-node3/TASK: Gather current interface info [028d2410-947f-82dc-c122-0000000000ec] 13731 1727203827.51613: sending task result for task 028d2410-947f-82dc-c122-0000000000ec 13731 1727203827.51712: done sending task result for task 028d2410-947f-82dc-c122-0000000000ec 13731 1727203827.51715: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": [ "ls", "-1" ], "delta": "0:00:00.003355", "end": "2024-09-24 14:50:27.473623", "rc": 0, "start": "2024-09-24 14:50:27.470268" } STDOUT: bonding_masters eth0 lo 13731 1727203827.51784: no more pending results, returning what we have 13731 1727203827.51788: results queue empty 13731 1727203827.51789: checking for any_errors_fatal 13731 1727203827.51790: done checking for any_errors_fatal 13731 1727203827.51791: checking for max_fail_percentage 13731 1727203827.51792: done checking for max_fail_percentage 13731 1727203827.51793: checking to see if all hosts have failed and the running result is not ok 13731 1727203827.51793: done checking to see if all hosts have failed 13731 1727203827.51794: getting the remaining hosts for this loop 13731 1727203827.51796: done getting the remaining hosts for this loop 13731 1727203827.51799: getting the next task for host managed-node3 13731 1727203827.51805: done getting next task for host managed-node3 13731 1727203827.51808: ^ task is: TASK: Set current_interfaces 13731 1727203827.51812: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203827.51817: getting variables 13731 1727203827.51818: in VariableManager get_vars() 13731 1727203827.51846: Calling all_inventory to load vars for managed-node3 13731 1727203827.51849: Calling groups_inventory to load vars for managed-node3 13731 1727203827.51852: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203827.51866: Calling all_plugins_play to load vars for managed-node3 13731 1727203827.51868: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203827.51871: Calling groups_plugins_play to load vars for managed-node3 13731 1727203827.52030: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203827.52173: done with get_vars() 13731 1727203827.52183: done getting variables 13731 1727203827.52240: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set current_interfaces] ************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml:9 Tuesday 24 September 2024 14:50:27 -0400 (0:00:00.620) 0:00:05.753 ***** 13731 1727203827.52264: entering _queue_task() for managed-node3/set_fact 13731 1727203827.52496: worker is 1 (out of 1 available) 13731 1727203827.52509: exiting _queue_task() for managed-node3/set_fact 13731 1727203827.52522: done queuing things up, now waiting for results queue to drain 13731 1727203827.52524: waiting for pending results... 13731 1727203827.52890: running TaskExecutor() for managed-node3/TASK: Set current_interfaces 13731 1727203827.52896: in run() - task 028d2410-947f-82dc-c122-0000000000ed 13731 1727203827.52914: variable 'ansible_search_path' from source: unknown 13731 1727203827.52922: variable 'ansible_search_path' from source: unknown 13731 1727203827.52965: calling self._execute() 13731 1727203827.53046: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.53059: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.53079: variable 'omit' from source: magic vars 13731 1727203827.53469: variable 'ansible_distribution_major_version' from source: facts 13731 1727203827.53489: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203827.53500: variable 'omit' from source: magic vars 13731 1727203827.53554: variable 'omit' from source: magic vars 13731 1727203827.53677: variable '_current_interfaces' from source: set_fact 13731 1727203827.53836: variable 'omit' from source: magic vars 13731 1727203827.53839: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203827.53842: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203827.53845: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203827.53860: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203827.53873: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203827.53899: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203827.53902: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.53904: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.53999: Set connection var ansible_pipelining to False 13731 1727203827.54003: Set connection var ansible_shell_type to sh 13731 1727203827.54005: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203827.54011: Set connection var ansible_connection to ssh 13731 1727203827.54016: Set connection var ansible_shell_executable to /bin/sh 13731 1727203827.54021: Set connection var ansible_timeout to 10 13731 1727203827.54039: variable 'ansible_shell_executable' from source: unknown 13731 1727203827.54043: variable 'ansible_connection' from source: unknown 13731 1727203827.54045: variable 'ansible_module_compression' from source: unknown 13731 1727203827.54048: variable 'ansible_shell_type' from source: unknown 13731 1727203827.54050: variable 'ansible_shell_executable' from source: unknown 13731 1727203827.54052: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.54054: variable 'ansible_pipelining' from source: unknown 13731 1727203827.54057: variable 'ansible_timeout' from source: unknown 13731 1727203827.54061: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.54179: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203827.54188: variable 'omit' from source: magic vars 13731 1727203827.54193: starting attempt loop 13731 1727203827.54196: running the handler 13731 1727203827.54205: handler run complete 13731 1727203827.54214: attempt loop complete, returning result 13731 1727203827.54217: _execute() done 13731 1727203827.54219: dumping result to json 13731 1727203827.54222: done dumping result, returning 13731 1727203827.54229: done running TaskExecutor() for managed-node3/TASK: Set current_interfaces [028d2410-947f-82dc-c122-0000000000ed] 13731 1727203827.54231: sending task result for task 028d2410-947f-82dc-c122-0000000000ed 13731 1727203827.54311: done sending task result for task 028d2410-947f-82dc-c122-0000000000ed 13731 1727203827.54313: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "current_interfaces": [ "bonding_masters", "eth0", "lo" ] }, "changed": false } 13731 1727203827.54369: no more pending results, returning what we have 13731 1727203827.54372: results queue empty 13731 1727203827.54373: checking for any_errors_fatal 13731 1727203827.54385: done checking for any_errors_fatal 13731 1727203827.54386: checking for max_fail_percentage 13731 1727203827.54387: done checking for max_fail_percentage 13731 1727203827.54388: checking to see if all hosts have failed and the running result is not ok 13731 1727203827.54388: done checking to see if all hosts have failed 13731 1727203827.54389: getting the remaining hosts for this loop 13731 1727203827.54391: done getting the remaining hosts for this loop 13731 1727203827.54394: getting the next task for host managed-node3 13731 1727203827.54402: done getting next task for host managed-node3 13731 1727203827.54404: ^ task is: TASK: Show current_interfaces 13731 1727203827.54407: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203827.54410: getting variables 13731 1727203827.54411: in VariableManager get_vars() 13731 1727203827.54435: Calling all_inventory to load vars for managed-node3 13731 1727203827.54438: Calling groups_inventory to load vars for managed-node3 13731 1727203827.54440: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203827.54448: Calling all_plugins_play to load vars for managed-node3 13731 1727203827.54450: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203827.54452: Calling groups_plugins_play to load vars for managed-node3 13731 1727203827.54582: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203827.54702: done with get_vars() 13731 1727203827.54709: done getting variables 13731 1727203827.54747: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Show current_interfaces] ************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml:5 Tuesday 24 September 2024 14:50:27 -0400 (0:00:00.025) 0:00:05.778 ***** 13731 1727203827.54767: entering _queue_task() for managed-node3/debug 13731 1727203827.54956: worker is 1 (out of 1 available) 13731 1727203827.54969: exiting _queue_task() for managed-node3/debug 13731 1727203827.54982: done queuing things up, now waiting for results queue to drain 13731 1727203827.54984: waiting for pending results... 13731 1727203827.55132: running TaskExecutor() for managed-node3/TASK: Show current_interfaces 13731 1727203827.55202: in run() - task 028d2410-947f-82dc-c122-0000000000b2 13731 1727203827.55215: variable 'ansible_search_path' from source: unknown 13731 1727203827.55218: variable 'ansible_search_path' from source: unknown 13731 1727203827.55246: calling self._execute() 13731 1727203827.55300: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.55309: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.55319: variable 'omit' from source: magic vars 13731 1727203827.55616: variable 'ansible_distribution_major_version' from source: facts 13731 1727203827.55625: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203827.55636: variable 'omit' from source: magic vars 13731 1727203827.55663: variable 'omit' from source: magic vars 13731 1727203827.55750: variable 'current_interfaces' from source: set_fact 13731 1727203827.55820: variable 'omit' from source: magic vars 13731 1727203827.55823: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203827.55856: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203827.55899: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203827.55902: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203827.55905: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203827.55925: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203827.55927: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.55934: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.56028: Set connection var ansible_pipelining to False 13731 1727203827.56280: Set connection var ansible_shell_type to sh 13731 1727203827.56283: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203827.56285: Set connection var ansible_connection to ssh 13731 1727203827.56287: Set connection var ansible_shell_executable to /bin/sh 13731 1727203827.56292: Set connection var ansible_timeout to 10 13731 1727203827.56294: variable 'ansible_shell_executable' from source: unknown 13731 1727203827.56296: variable 'ansible_connection' from source: unknown 13731 1727203827.56299: variable 'ansible_module_compression' from source: unknown 13731 1727203827.56300: variable 'ansible_shell_type' from source: unknown 13731 1727203827.56302: variable 'ansible_shell_executable' from source: unknown 13731 1727203827.56304: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.56306: variable 'ansible_pipelining' from source: unknown 13731 1727203827.56308: variable 'ansible_timeout' from source: unknown 13731 1727203827.56309: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.56312: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203827.56314: variable 'omit' from source: magic vars 13731 1727203827.56315: starting attempt loop 13731 1727203827.56317: running the handler 13731 1727203827.56342: handler run complete 13731 1727203827.56360: attempt loop complete, returning result 13731 1727203827.56374: _execute() done 13731 1727203827.56383: dumping result to json 13731 1727203827.56389: done dumping result, returning 13731 1727203827.56398: done running TaskExecutor() for managed-node3/TASK: Show current_interfaces [028d2410-947f-82dc-c122-0000000000b2] 13731 1727203827.56407: sending task result for task 028d2410-947f-82dc-c122-0000000000b2 13731 1727203827.56511: done sending task result for task 028d2410-947f-82dc-c122-0000000000b2 13731 1727203827.56513: WORKER PROCESS EXITING ok: [managed-node3] => {} MSG: current_interfaces: ['bonding_masters', 'eth0', 'lo'] 13731 1727203827.56587: no more pending results, returning what we have 13731 1727203827.56591: results queue empty 13731 1727203827.56592: checking for any_errors_fatal 13731 1727203827.56598: done checking for any_errors_fatal 13731 1727203827.56598: checking for max_fail_percentage 13731 1727203827.56600: done checking for max_fail_percentage 13731 1727203827.56600: checking to see if all hosts have failed and the running result is not ok 13731 1727203827.56601: done checking to see if all hosts have failed 13731 1727203827.56601: getting the remaining hosts for this loop 13731 1727203827.56603: done getting the remaining hosts for this loop 13731 1727203827.56609: getting the next task for host managed-node3 13731 1727203827.56617: done getting next task for host managed-node3 13731 1727203827.56619: ^ task is: TASK: Setup 13731 1727203827.56621: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203827.56625: getting variables 13731 1727203827.56626: in VariableManager get_vars() 13731 1727203827.56650: Calling all_inventory to load vars for managed-node3 13731 1727203827.56653: Calling groups_inventory to load vars for managed-node3 13731 1727203827.56656: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203827.56666: Calling all_plugins_play to load vars for managed-node3 13731 1727203827.56669: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203827.56671: Calling groups_plugins_play to load vars for managed-node3 13731 1727203827.56852: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203827.57027: done with get_vars() 13731 1727203827.57039: done getting variables TASK [Setup] ******************************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:24 Tuesday 24 September 2024 14:50:27 -0400 (0:00:00.023) 0:00:05.802 ***** 13731 1727203827.57129: entering _queue_task() for managed-node3/include_tasks 13731 1727203827.57369: worker is 1 (out of 1 available) 13731 1727203827.57382: exiting _queue_task() for managed-node3/include_tasks 13731 1727203827.57394: done queuing things up, now waiting for results queue to drain 13731 1727203827.57395: waiting for pending results... 13731 1727203827.57795: running TaskExecutor() for managed-node3/TASK: Setup 13731 1727203827.57800: in run() - task 028d2410-947f-82dc-c122-00000000008b 13731 1727203827.57803: variable 'ansible_search_path' from source: unknown 13731 1727203827.57806: variable 'ansible_search_path' from source: unknown 13731 1727203827.57809: variable 'lsr_setup' from source: include params 13731 1727203827.58016: variable 'lsr_setup' from source: include params 13731 1727203827.58073: variable 'omit' from source: magic vars 13731 1727203827.58153: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.58160: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.58173: variable 'omit' from source: magic vars 13731 1727203827.58340: variable 'ansible_distribution_major_version' from source: facts 13731 1727203827.58350: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203827.58353: variable 'item' from source: unknown 13731 1727203827.58403: variable 'item' from source: unknown 13731 1727203827.58426: variable 'item' from source: unknown 13731 1727203827.58472: variable 'item' from source: unknown 13731 1727203827.58578: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.58581: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.58584: variable 'omit' from source: magic vars 13731 1727203827.58663: variable 'ansible_distribution_major_version' from source: facts 13731 1727203827.58670: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203827.58677: variable 'item' from source: unknown 13731 1727203827.58722: variable 'item' from source: unknown 13731 1727203827.58742: variable 'item' from source: unknown 13731 1727203827.58787: variable 'item' from source: unknown 13731 1727203827.58846: dumping result to json 13731 1727203827.58849: done dumping result, returning 13731 1727203827.58851: done running TaskExecutor() for managed-node3/TASK: Setup [028d2410-947f-82dc-c122-00000000008b] 13731 1727203827.58853: sending task result for task 028d2410-947f-82dc-c122-00000000008b 13731 1727203827.58886: done sending task result for task 028d2410-947f-82dc-c122-00000000008b 13731 1727203827.58889: WORKER PROCESS EXITING 13731 1727203827.58912: no more pending results, returning what we have 13731 1727203827.58917: in VariableManager get_vars() 13731 1727203827.58948: Calling all_inventory to load vars for managed-node3 13731 1727203827.58950: Calling groups_inventory to load vars for managed-node3 13731 1727203827.58954: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203827.58965: Calling all_plugins_play to load vars for managed-node3 13731 1727203827.58968: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203827.58970: Calling groups_plugins_play to load vars for managed-node3 13731 1727203827.59113: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203827.59223: done with get_vars() 13731 1727203827.59228: variable 'ansible_search_path' from source: unknown 13731 1727203827.59229: variable 'ansible_search_path' from source: unknown 13731 1727203827.59254: variable 'ansible_search_path' from source: unknown 13731 1727203827.59255: variable 'ansible_search_path' from source: unknown 13731 1727203827.59271: we have included files to process 13731 1727203827.59272: generating all_blocks data 13731 1727203827.59273: done generating all_blocks data 13731 1727203827.59278: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml 13731 1727203827.59279: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml 13731 1727203827.59281: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml 13731 1727203827.60034: done processing included file 13731 1727203827.60036: iterating over new_blocks loaded from include file 13731 1727203827.60038: in VariableManager get_vars() 13731 1727203827.60051: done with get_vars() 13731 1727203827.60052: filtering new block on tags 13731 1727203827.60124: done filtering new block on tags 13731 1727203827.60127: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml for managed-node3 => (item=tasks/create_test_interfaces_with_dhcp.yml) 13731 1727203827.60132: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_dhcp_device_present.yml 13731 1727203827.60134: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_dhcp_device_present.yml 13731 1727203827.60137: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_dhcp_device_present.yml 13731 1727203827.60257: in VariableManager get_vars() 13731 1727203827.60273: done with get_vars() 13731 1727203827.60281: variable 'item' from source: include params 13731 1727203827.60388: variable 'item' from source: include params 13731 1727203827.60421: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml statically imported: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml 13731 1727203827.60530: in VariableManager get_vars() 13731 1727203827.60549: done with get_vars() 13731 1727203827.60750: in VariableManager get_vars() 13731 1727203827.60768: done with get_vars() 13731 1727203827.60774: variable 'item' from source: include params 13731 1727203827.60857: variable 'item' from source: include params 13731 1727203827.60900: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml statically imported: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml 13731 1727203827.60978: in VariableManager get_vars() 13731 1727203827.61011: done with get_vars() 13731 1727203827.61091: done processing included file 13731 1727203827.61092: iterating over new_blocks loaded from include file 13731 1727203827.61093: in VariableManager get_vars() 13731 1727203827.61102: done with get_vars() 13731 1727203827.61103: filtering new block on tags 13731 1727203827.61158: done filtering new block on tags 13731 1727203827.61161: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_dhcp_device_present.yml for managed-node3 => (item=tasks/assert_dhcp_device_present.yml) 13731 1727203827.61164: extending task lists for all hosts with included blocks 13731 1727203827.61524: done extending task lists 13731 1727203827.61525: done processing included files 13731 1727203827.61526: results queue empty 13731 1727203827.61526: checking for any_errors_fatal 13731 1727203827.61529: done checking for any_errors_fatal 13731 1727203827.61529: checking for max_fail_percentage 13731 1727203827.61530: done checking for max_fail_percentage 13731 1727203827.61530: checking to see if all hosts have failed and the running result is not ok 13731 1727203827.61531: done checking to see if all hosts have failed 13731 1727203827.61535: getting the remaining hosts for this loop 13731 1727203827.61536: done getting the remaining hosts for this loop 13731 1727203827.61538: getting the next task for host managed-node3 13731 1727203827.61540: done getting next task for host managed-node3 13731 1727203827.61542: ^ task is: TASK: Install dnsmasq 13731 1727203827.61544: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203827.61545: getting variables 13731 1727203827.61546: in VariableManager get_vars() 13731 1727203827.61551: Calling all_inventory to load vars for managed-node3 13731 1727203827.61553: Calling groups_inventory to load vars for managed-node3 13731 1727203827.61554: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203827.61558: Calling all_plugins_play to load vars for managed-node3 13731 1727203827.61560: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203827.61561: Calling groups_plugins_play to load vars for managed-node3 13731 1727203827.61657: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203827.61765: done with get_vars() 13731 1727203827.61771: done getting variables 13731 1727203827.61798: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Install dnsmasq] ********************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:3 Tuesday 24 September 2024 14:50:27 -0400 (0:00:00.046) 0:00:05.849 ***** 13731 1727203827.61816: entering _queue_task() for managed-node3/package 13731 1727203827.62042: worker is 1 (out of 1 available) 13731 1727203827.62059: exiting _queue_task() for managed-node3/package 13731 1727203827.62071: done queuing things up, now waiting for results queue to drain 13731 1727203827.62073: waiting for pending results... 13731 1727203827.62235: running TaskExecutor() for managed-node3/TASK: Install dnsmasq 13731 1727203827.62299: in run() - task 028d2410-947f-82dc-c122-000000000112 13731 1727203827.62311: variable 'ansible_search_path' from source: unknown 13731 1727203827.62315: variable 'ansible_search_path' from source: unknown 13731 1727203827.62340: calling self._execute() 13731 1727203827.62444: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.62450: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.62458: variable 'omit' from source: magic vars 13731 1727203827.62720: variable 'ansible_distribution_major_version' from source: facts 13731 1727203827.62730: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203827.62738: variable 'omit' from source: magic vars 13731 1727203827.62768: variable 'omit' from source: magic vars 13731 1727203827.63003: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203827.65216: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203827.65287: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203827.65345: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203827.65389: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203827.65431: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203827.65645: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203827.65649: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203827.65653: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203827.65682: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203827.65704: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203827.65832: variable '__network_is_ostree' from source: set_fact 13731 1727203827.65843: variable 'omit' from source: magic vars 13731 1727203827.65892: variable 'omit' from source: magic vars 13731 1727203827.65926: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203827.65972: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203827.65998: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203827.66083: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203827.66086: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203827.66088: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203827.66090: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.66094: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.66381: Set connection var ansible_pipelining to False 13731 1727203827.66384: Set connection var ansible_shell_type to sh 13731 1727203827.66387: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203827.66389: Set connection var ansible_connection to ssh 13731 1727203827.66391: Set connection var ansible_shell_executable to /bin/sh 13731 1727203827.66393: Set connection var ansible_timeout to 10 13731 1727203827.66395: variable 'ansible_shell_executable' from source: unknown 13731 1727203827.66397: variable 'ansible_connection' from source: unknown 13731 1727203827.66399: variable 'ansible_module_compression' from source: unknown 13731 1727203827.66400: variable 'ansible_shell_type' from source: unknown 13731 1727203827.66402: variable 'ansible_shell_executable' from source: unknown 13731 1727203827.66404: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203827.66406: variable 'ansible_pipelining' from source: unknown 13731 1727203827.66408: variable 'ansible_timeout' from source: unknown 13731 1727203827.66410: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203827.66422: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203827.66438: variable 'omit' from source: magic vars 13731 1727203827.66448: starting attempt loop 13731 1727203827.66454: running the handler 13731 1727203827.66468: variable 'ansible_facts' from source: unknown 13731 1727203827.66529: variable 'ansible_facts' from source: unknown 13731 1727203827.66541: _low_level_execute_command(): starting 13731 1727203827.66554: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203827.67408: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.67452: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203827.67486: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203827.67502: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.67587: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.69250: stdout chunk (state=3): >>>/root <<< 13731 1727203827.69377: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203827.69414: stderr chunk (state=3): >>><<< 13731 1727203827.69418: stdout chunk (state=3): >>><<< 13731 1727203827.69484: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203827.69497: _low_level_execute_command(): starting 13731 1727203827.69500: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214 `" && echo ansible-tmp-1727203827.6943843-14143-94690930111214="` echo /root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214 `" ) && sleep 0' 13731 1727203827.70130: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203827.70147: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203827.70168: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203827.70191: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203827.70251: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.70313: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203827.70331: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203827.70368: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.70436: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.72318: stdout chunk (state=3): >>>ansible-tmp-1727203827.6943843-14143-94690930111214=/root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214 <<< 13731 1727203827.72479: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203827.72483: stdout chunk (state=3): >>><<< 13731 1727203827.72486: stderr chunk (state=3): >>><<< 13731 1727203827.72525: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203827.6943843-14143-94690930111214=/root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203827.72586: variable 'ansible_module_compression' from source: unknown 13731 1727203827.72645: ANSIBALLZ: Using generic lock for ansible.legacy.dnf 13731 1727203827.72681: ANSIBALLZ: Acquiring lock 13731 1727203827.72685: ANSIBALLZ: Lock acquired: 140078454804688 13731 1727203827.72687: ANSIBALLZ: Creating module 13731 1727203827.84823: ANSIBALLZ: Writing module into payload 13731 1727203827.85013: ANSIBALLZ: Writing module 13731 1727203827.85247: ANSIBALLZ: Renaming module 13731 1727203827.85250: ANSIBALLZ: Done creating module 13731 1727203827.85252: variable 'ansible_facts' from source: unknown 13731 1727203827.85254: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/AnsiballZ_dnf.py 13731 1727203827.85509: Sending initial data 13731 1727203827.85513: Sent initial data (151 bytes) 13731 1727203827.86708: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203827.86724: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.86912: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203827.86925: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.87004: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.88666: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203827.88684: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203827.88733: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpm4u1dw_y /root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/AnsiballZ_dnf.py <<< 13731 1727203827.88736: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/AnsiballZ_dnf.py" <<< 13731 1727203827.88798: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpm4u1dw_y" to remote "/root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/AnsiballZ_dnf.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/AnsiballZ_dnf.py" <<< 13731 1727203827.89800: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203827.89852: stderr chunk (state=3): >>><<< 13731 1727203827.89882: stdout chunk (state=3): >>><<< 13731 1727203827.89907: done transferring module to remote 13731 1727203827.89924: _low_level_execute_command(): starting 13731 1727203827.89991: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/ /root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/AnsiballZ_dnf.py && sleep 0' 13731 1727203827.90649: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203827.90681: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203827.90765: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.90805: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203827.90829: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203827.90842: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.90912: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203827.92699: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203827.92703: stdout chunk (state=3): >>><<< 13731 1727203827.92709: stderr chunk (state=3): >>><<< 13731 1727203827.92736: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203827.92826: _low_level_execute_command(): starting 13731 1727203827.92830: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/AnsiballZ_dnf.py && sleep 0' 13731 1727203827.93397: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203827.93414: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203827.93433: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203827.93451: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203827.93472: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203827.93487: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203827.93541: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203827.93597: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203827.93614: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203827.93637: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203827.93711: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203828.34112: stdout chunk (state=3): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["dnsmasq"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} <<< 13731 1727203828.38469: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203828.38474: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203828.38479: stdout chunk (state=3): >>><<< 13731 1727203828.38482: stderr chunk (state=3): >>><<< 13731 1727203828.38485: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["dnsmasq"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203828.38491: done with _execute_module (ansible.legacy.dnf, {'name': 'dnsmasq', 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203828.38493: _low_level_execute_command(): starting 13731 1727203828.38496: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203827.6943843-14143-94690930111214/ > /dev/null 2>&1 && sleep 0' 13731 1727203828.39480: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203828.39484: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203828.39583: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203828.39687: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203828.41517: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203828.41541: stderr chunk (state=3): >>><<< 13731 1727203828.41615: stdout chunk (state=3): >>><<< 13731 1727203828.41843: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203828.41846: handler run complete 13731 1727203828.42070: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203828.42260: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203828.42314: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203828.42351: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203828.42394: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203828.42469: variable '__install_status' from source: unknown 13731 1727203828.42503: Evaluated conditional (__install_status is success): True 13731 1727203828.42523: attempt loop complete, returning result 13731 1727203828.42530: _execute() done 13731 1727203828.42536: dumping result to json 13731 1727203828.42544: done dumping result, returning 13731 1727203828.42555: done running TaskExecutor() for managed-node3/TASK: Install dnsmasq [028d2410-947f-82dc-c122-000000000112] 13731 1727203828.42566: sending task result for task 028d2410-947f-82dc-c122-000000000112 ok: [managed-node3] => { "attempts": 1, "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 13731 1727203828.42820: no more pending results, returning what we have 13731 1727203828.42824: results queue empty 13731 1727203828.42825: checking for any_errors_fatal 13731 1727203828.42827: done checking for any_errors_fatal 13731 1727203828.42827: checking for max_fail_percentage 13731 1727203828.42829: done checking for max_fail_percentage 13731 1727203828.42830: checking to see if all hosts have failed and the running result is not ok 13731 1727203828.42830: done checking to see if all hosts have failed 13731 1727203828.42831: getting the remaining hosts for this loop 13731 1727203828.42833: done getting the remaining hosts for this loop 13731 1727203828.42836: getting the next task for host managed-node3 13731 1727203828.42843: done getting next task for host managed-node3 13731 1727203828.42845: ^ task is: TASK: Install pgrep, sysctl 13731 1727203828.42849: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203828.42852: getting variables 13731 1727203828.42854: in VariableManager get_vars() 13731 1727203828.42885: Calling all_inventory to load vars for managed-node3 13731 1727203828.42889: Calling groups_inventory to load vars for managed-node3 13731 1727203828.42893: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203828.42906: Calling all_plugins_play to load vars for managed-node3 13731 1727203828.42910: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203828.42913: Calling groups_plugins_play to load vars for managed-node3 13731 1727203828.43414: done sending task result for task 028d2410-947f-82dc-c122-000000000112 13731 1727203828.43417: WORKER PROCESS EXITING 13731 1727203828.43430: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203828.43727: done with get_vars() 13731 1727203828.43742: done getting variables 13731 1727203828.43803: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Install pgrep, sysctl] *************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:17 Tuesday 24 September 2024 14:50:28 -0400 (0:00:00.820) 0:00:06.669 ***** 13731 1727203828.43837: entering _queue_task() for managed-node3/package 13731 1727203828.44203: worker is 1 (out of 1 available) 13731 1727203828.44213: exiting _queue_task() for managed-node3/package 13731 1727203828.44225: done queuing things up, now waiting for results queue to drain 13731 1727203828.44226: waiting for pending results... 13731 1727203828.44389: running TaskExecutor() for managed-node3/TASK: Install pgrep, sysctl 13731 1727203828.44507: in run() - task 028d2410-947f-82dc-c122-000000000113 13731 1727203828.44525: variable 'ansible_search_path' from source: unknown 13731 1727203828.44532: variable 'ansible_search_path' from source: unknown 13731 1727203828.44574: calling self._execute() 13731 1727203828.44655: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203828.44674: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203828.44689: variable 'omit' from source: magic vars 13731 1727203828.45074: variable 'ansible_distribution_major_version' from source: facts 13731 1727203828.45094: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203828.45261: variable 'ansible_os_family' from source: facts 13731 1727203828.45267: Evaluated conditional (ansible_os_family == 'RedHat'): True 13731 1727203828.45401: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203828.45745: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203828.45809: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203828.45916: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203828.45919: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203828.46005: variable 'ansible_distribution_major_version' from source: facts 13731 1727203828.46048: Evaluated conditional (ansible_distribution_major_version is version('6', '<=')): False 13731 1727203828.46057: when evaluation is False, skipping this task 13731 1727203828.46068: _execute() done 13731 1727203828.46074: dumping result to json 13731 1727203828.46084: done dumping result, returning 13731 1727203828.46133: done running TaskExecutor() for managed-node3/TASK: Install pgrep, sysctl [028d2410-947f-82dc-c122-000000000113] 13731 1727203828.46137: sending task result for task 028d2410-947f-82dc-c122-000000000113 skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version is version('6', '<=')", "skip_reason": "Conditional result was False" } 13731 1727203828.46268: no more pending results, returning what we have 13731 1727203828.46273: results queue empty 13731 1727203828.46274: checking for any_errors_fatal 13731 1727203828.46283: done checking for any_errors_fatal 13731 1727203828.46284: checking for max_fail_percentage 13731 1727203828.46286: done checking for max_fail_percentage 13731 1727203828.46286: checking to see if all hosts have failed and the running result is not ok 13731 1727203828.46287: done checking to see if all hosts have failed 13731 1727203828.46288: getting the remaining hosts for this loop 13731 1727203828.46290: done getting the remaining hosts for this loop 13731 1727203828.46294: getting the next task for host managed-node3 13731 1727203828.46301: done getting next task for host managed-node3 13731 1727203828.46305: ^ task is: TASK: Install pgrep, sysctl 13731 1727203828.46308: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203828.46312: getting variables 13731 1727203828.46314: in VariableManager get_vars() 13731 1727203828.46474: Calling all_inventory to load vars for managed-node3 13731 1727203828.46480: Calling groups_inventory to load vars for managed-node3 13731 1727203828.46484: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203828.46496: Calling all_plugins_play to load vars for managed-node3 13731 1727203828.46499: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203828.46503: Calling groups_plugins_play to load vars for managed-node3 13731 1727203828.46577: done sending task result for task 028d2410-947f-82dc-c122-000000000113 13731 1727203828.46581: WORKER PROCESS EXITING 13731 1727203828.46985: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203828.47234: done with get_vars() 13731 1727203828.47244: done getting variables 13731 1727203828.47304: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Install pgrep, sysctl] *************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:26 Tuesday 24 September 2024 14:50:28 -0400 (0:00:00.034) 0:00:06.704 ***** 13731 1727203828.47333: entering _queue_task() for managed-node3/package 13731 1727203828.47704: worker is 1 (out of 1 available) 13731 1727203828.47715: exiting _queue_task() for managed-node3/package 13731 1727203828.47725: done queuing things up, now waiting for results queue to drain 13731 1727203828.47727: waiting for pending results... 13731 1727203828.47884: running TaskExecutor() for managed-node3/TASK: Install pgrep, sysctl 13731 1727203828.47998: in run() - task 028d2410-947f-82dc-c122-000000000114 13731 1727203828.48061: variable 'ansible_search_path' from source: unknown 13731 1727203828.48066: variable 'ansible_search_path' from source: unknown 13731 1727203828.48069: calling self._execute() 13731 1727203828.48143: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203828.48153: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203828.48176: variable 'omit' from source: magic vars 13731 1727203828.48549: variable 'ansible_distribution_major_version' from source: facts 13731 1727203828.48568: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203828.48713: variable 'ansible_os_family' from source: facts 13731 1727203828.48716: Evaluated conditional (ansible_os_family == 'RedHat'): True 13731 1727203828.48888: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203828.49171: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203828.49290: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203828.49294: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203828.49296: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203828.49373: variable 'ansible_distribution_major_version' from source: facts 13731 1727203828.49397: Evaluated conditional (ansible_distribution_major_version is version('7', '>=')): True 13731 1727203828.49413: variable 'omit' from source: magic vars 13731 1727203828.49457: variable 'omit' from source: magic vars 13731 1727203828.49627: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203828.52142: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203828.52251: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203828.52256: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203828.52292: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203828.52322: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203828.52425: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203828.52480: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203828.52509: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203828.52580: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203828.52587: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203828.52696: variable '__network_is_ostree' from source: set_fact 13731 1727203828.52791: variable 'omit' from source: magic vars 13731 1727203828.52794: variable 'omit' from source: magic vars 13731 1727203828.52796: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203828.52798: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203828.52824: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203828.52844: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203828.52858: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203828.52900: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203828.52908: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203828.52919: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203828.53031: Set connection var ansible_pipelining to False 13731 1727203828.53044: Set connection var ansible_shell_type to sh 13731 1727203828.53056: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203828.53069: Set connection var ansible_connection to ssh 13731 1727203828.53080: Set connection var ansible_shell_executable to /bin/sh 13731 1727203828.53089: Set connection var ansible_timeout to 10 13731 1727203828.53142: variable 'ansible_shell_executable' from source: unknown 13731 1727203828.53151: variable 'ansible_connection' from source: unknown 13731 1727203828.53158: variable 'ansible_module_compression' from source: unknown 13731 1727203828.53225: variable 'ansible_shell_type' from source: unknown 13731 1727203828.53229: variable 'ansible_shell_executable' from source: unknown 13731 1727203828.53231: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203828.53233: variable 'ansible_pipelining' from source: unknown 13731 1727203828.53235: variable 'ansible_timeout' from source: unknown 13731 1727203828.53237: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203828.53306: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203828.53323: variable 'omit' from source: magic vars 13731 1727203828.53338: starting attempt loop 13731 1727203828.53345: running the handler 13731 1727203828.53361: variable 'ansible_facts' from source: unknown 13731 1727203828.53370: variable 'ansible_facts' from source: unknown 13731 1727203828.53425: _low_level_execute_command(): starting 13731 1727203828.53443: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203828.54187: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203828.54206: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203828.54223: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203828.54330: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203828.54355: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203828.54378: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203828.54400: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203828.54471: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203828.56305: stdout chunk (state=3): >>>/root <<< 13731 1727203828.56389: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203828.56393: stdout chunk (state=3): >>><<< 13731 1727203828.56395: stderr chunk (state=3): >>><<< 13731 1727203828.56414: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203828.56442: _low_level_execute_command(): starting 13731 1727203828.56454: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824 `" && echo ansible-tmp-1727203828.5642936-14333-142659790261824="` echo /root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824 `" ) && sleep 0' 13731 1727203828.57110: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203828.57131: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203828.57145: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203828.57160: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203828.57184: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203828.57242: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203828.57297: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203828.57320: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203828.57342: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203828.57411: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203828.59323: stdout chunk (state=3): >>>ansible-tmp-1727203828.5642936-14333-142659790261824=/root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824 <<< 13731 1727203828.59505: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203828.59509: stdout chunk (state=3): >>><<< 13731 1727203828.59511: stderr chunk (state=3): >>><<< 13731 1727203828.59534: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203828.5642936-14333-142659790261824=/root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203828.59682: variable 'ansible_module_compression' from source: unknown 13731 1727203828.59685: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.dnf-ZIP_DEFLATED 13731 1727203828.59697: variable 'ansible_facts' from source: unknown 13731 1727203828.59830: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/AnsiballZ_dnf.py 13731 1727203828.59996: Sending initial data 13731 1727203828.60142: Sent initial data (152 bytes) 13731 1727203828.60679: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203828.60790: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203828.60865: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203828.60906: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203828.62419: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203828.62458: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203828.62517: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpfql8m6s4 /root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/AnsiballZ_dnf.py <<< 13731 1727203828.62520: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/AnsiballZ_dnf.py" <<< 13731 1727203828.62555: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpfql8m6s4" to remote "/root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/AnsiballZ_dnf.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/AnsiballZ_dnf.py" <<< 13731 1727203828.63482: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203828.63654: stderr chunk (state=3): >>><<< 13731 1727203828.63657: stdout chunk (state=3): >>><<< 13731 1727203828.63659: done transferring module to remote 13731 1727203828.63661: _low_level_execute_command(): starting 13731 1727203828.63663: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/ /root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/AnsiballZ_dnf.py && sleep 0' 13731 1727203828.64238: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203828.64255: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203828.64272: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203828.64329: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203828.64340: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203828.64431: stderr chunk (state=3): >>>debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203828.64455: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203828.64541: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203828.66300: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203828.66327: stderr chunk (state=3): >>><<< 13731 1727203828.66340: stdout chunk (state=3): >>><<< 13731 1727203828.66446: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203828.66449: _low_level_execute_command(): starting 13731 1727203828.66452: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/AnsiballZ_dnf.py && sleep 0' 13731 1727203828.67021: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203828.67035: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203828.67091: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203828.67160: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203828.67181: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203828.67202: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203828.67266: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203829.07292: stdout chunk (state=3): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["procps-ng"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} <<< 13731 1727203829.11788: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203829.11792: stdout chunk (state=3): >>><<< 13731 1727203829.11795: stderr chunk (state=3): >>><<< 13731 1727203829.11797: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["procps-ng"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203829.11804: done with _execute_module (ansible.legacy.dnf, {'name': 'procps-ng', 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203829.11806: _low_level_execute_command(): starting 13731 1727203829.11808: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203828.5642936-14333-142659790261824/ > /dev/null 2>&1 && sleep 0' 13731 1727203829.12718: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203829.12761: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203829.12781: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203829.12903: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203829.12920: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203829.12935: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203829.12956: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203829.13242: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203829.15049: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203829.15062: stdout chunk (state=3): >>><<< 13731 1727203829.15085: stderr chunk (state=3): >>><<< 13731 1727203829.15113: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203829.15127: handler run complete 13731 1727203829.15169: attempt loop complete, returning result 13731 1727203829.15179: _execute() done 13731 1727203829.15186: dumping result to json 13731 1727203829.15197: done dumping result, returning 13731 1727203829.15216: done running TaskExecutor() for managed-node3/TASK: Install pgrep, sysctl [028d2410-947f-82dc-c122-000000000114] 13731 1727203829.15225: sending task result for task 028d2410-947f-82dc-c122-000000000114 ok: [managed-node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 13731 1727203829.15558: no more pending results, returning what we have 13731 1727203829.15562: results queue empty 13731 1727203829.15563: checking for any_errors_fatal 13731 1727203829.15569: done checking for any_errors_fatal 13731 1727203829.15569: checking for max_fail_percentage 13731 1727203829.15571: done checking for max_fail_percentage 13731 1727203829.15572: checking to see if all hosts have failed and the running result is not ok 13731 1727203829.15572: done checking to see if all hosts have failed 13731 1727203829.15573: getting the remaining hosts for this loop 13731 1727203829.15630: done getting the remaining hosts for this loop 13731 1727203829.15635: getting the next task for host managed-node3 13731 1727203829.15642: done getting next task for host managed-node3 13731 1727203829.15645: ^ task is: TASK: Create test interfaces 13731 1727203829.15648: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203829.15651: getting variables 13731 1727203829.15653: in VariableManager get_vars() 13731 1727203829.15686: Calling all_inventory to load vars for managed-node3 13731 1727203829.15689: Calling groups_inventory to load vars for managed-node3 13731 1727203829.15693: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203829.16288: Calling all_plugins_play to load vars for managed-node3 13731 1727203829.16292: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203829.16297: Calling groups_plugins_play to load vars for managed-node3 13731 1727203829.16472: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203829.17196: done sending task result for task 028d2410-947f-82dc-c122-000000000114 13731 1727203829.17200: WORKER PROCESS EXITING 13731 1727203829.17469: done with get_vars() 13731 1727203829.17486: done getting variables 13731 1727203829.17655: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [Create test interfaces] ************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:35 Tuesday 24 September 2024 14:50:29 -0400 (0:00:00.704) 0:00:07.408 ***** 13731 1727203829.17745: entering _queue_task() for managed-node3/shell 13731 1727203829.17747: Creating lock for shell 13731 1727203829.18725: worker is 1 (out of 1 available) 13731 1727203829.18739: exiting _queue_task() for managed-node3/shell 13731 1727203829.18751: done queuing things up, now waiting for results queue to drain 13731 1727203829.18753: waiting for pending results... 13731 1727203829.19340: running TaskExecutor() for managed-node3/TASK: Create test interfaces 13731 1727203829.19451: in run() - task 028d2410-947f-82dc-c122-000000000115 13731 1727203829.19800: variable 'ansible_search_path' from source: unknown 13731 1727203829.19981: variable 'ansible_search_path' from source: unknown 13731 1727203829.19985: calling self._execute() 13731 1727203829.19987: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203829.19990: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203829.19992: variable 'omit' from source: magic vars 13731 1727203829.21100: variable 'ansible_distribution_major_version' from source: facts 13731 1727203829.21482: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203829.21486: variable 'omit' from source: magic vars 13731 1727203829.21488: variable 'omit' from source: magic vars 13731 1727203829.22720: variable 'dhcp_interface1' from source: play vars 13731 1727203829.22792: variable 'dhcp_interface2' from source: play vars 13731 1727203829.22824: variable 'omit' from source: magic vars 13731 1727203829.23027: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203829.23071: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203829.23101: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203829.23126: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203829.23142: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203829.23288: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203829.23298: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203829.23308: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203829.23682: Set connection var ansible_pipelining to False 13731 1727203829.23685: Set connection var ansible_shell_type to sh 13731 1727203829.23688: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203829.23690: Set connection var ansible_connection to ssh 13731 1727203829.23692: Set connection var ansible_shell_executable to /bin/sh 13731 1727203829.23782: Set connection var ansible_timeout to 10 13731 1727203829.23786: variable 'ansible_shell_executable' from source: unknown 13731 1727203829.23788: variable 'ansible_connection' from source: unknown 13731 1727203829.23790: variable 'ansible_module_compression' from source: unknown 13731 1727203829.23792: variable 'ansible_shell_type' from source: unknown 13731 1727203829.23794: variable 'ansible_shell_executable' from source: unknown 13731 1727203829.23796: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203829.23798: variable 'ansible_pipelining' from source: unknown 13731 1727203829.23800: variable 'ansible_timeout' from source: unknown 13731 1727203829.23802: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203829.23900: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203829.24094: variable 'omit' from source: magic vars 13731 1727203829.24104: starting attempt loop 13731 1727203829.24111: running the handler 13731 1727203829.24126: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203829.24153: _low_level_execute_command(): starting 13731 1727203829.24172: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203829.25808: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203829.25885: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203829.25902: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203829.26294: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203829.26365: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203829.27980: stdout chunk (state=3): >>>/root <<< 13731 1727203829.28111: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203829.28122: stdout chunk (state=3): >>><<< 13731 1727203829.28134: stderr chunk (state=3): >>><<< 13731 1727203829.28158: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203829.28485: _low_level_execute_command(): starting 13731 1727203829.28490: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334 `" && echo ansible-tmp-1727203829.2838345-14362-210838854669334="` echo /root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334 `" ) && sleep 0' 13731 1727203829.29551: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203829.29889: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203829.29920: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203829.29990: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203829.31840: stdout chunk (state=3): >>>ansible-tmp-1727203829.2838345-14362-210838854669334=/root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334 <<< 13731 1727203829.32092: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203829.32096: stdout chunk (state=3): >>><<< 13731 1727203829.32098: stderr chunk (state=3): >>><<< 13731 1727203829.32117: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203829.2838345-14362-210838854669334=/root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203829.32154: variable 'ansible_module_compression' from source: unknown 13731 1727203829.32218: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203829.32411: variable 'ansible_facts' from source: unknown 13731 1727203829.32494: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/AnsiballZ_command.py 13731 1727203829.33105: Sending initial data 13731 1727203829.33108: Sent initial data (156 bytes) 13731 1727203829.34156: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203829.34160: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203829.34288: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203829.34394: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203829.34444: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203829.34457: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203829.34509: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203829.36053: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203829.36084: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203829.36140: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpf60tvvcx /root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/AnsiballZ_command.py <<< 13731 1727203829.36152: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/AnsiballZ_command.py" <<< 13731 1727203829.36213: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpf60tvvcx" to remote "/root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/AnsiballZ_command.py" <<< 13731 1727203829.37282: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203829.37286: stdout chunk (state=3): >>><<< 13731 1727203829.37289: stderr chunk (state=3): >>><<< 13731 1727203829.37291: done transferring module to remote 13731 1727203829.37293: _low_level_execute_command(): starting 13731 1727203829.37295: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/ /root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/AnsiballZ_command.py && sleep 0' 13731 1727203829.38439: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203829.38489: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203829.38504: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203829.38525: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203829.38550: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203829.38589: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203829.38691: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203829.38703: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203829.38725: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203829.38792: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203829.40522: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203829.40617: stderr chunk (state=3): >>><<< 13731 1727203829.40625: stdout chunk (state=3): >>><<< 13731 1727203829.40660: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203829.40672: _low_level_execute_command(): starting 13731 1727203829.40685: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/AnsiballZ_command.py && sleep 0' 13731 1727203829.41350: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203829.41369: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203829.41387: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203829.41406: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203829.41430: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203829.41447: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203829.41496: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203829.41665: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203829.41805: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203829.41837: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203830.77839: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ ip link add test1 type veth peer name test1p\n+ ip link add test2 type veth peer name test2p\n++ pgrep NetworkManager\n+ '[' -n 704 ']'\n+ nmcli d set test1 managed true\n+ nmcli d set test2 managed true\n+ nmcli d set test1p managed false\n+ nmcli d set test2p managed false\n+ ip link set test1p up\n+ ip link set test2p up\n+ ip link add name testbr type bridge forward_delay 0\n++ pgrep NetworkManager\n+ '[' -n 704 ']'\n+ nmcli d set testbr managed false\n+ ip link set testbr up\n+ timer=0\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ let timer+=1\n+ '[' 1 -eq 30 ']'\n+ sleep 1\n+ rc=0\n+ ip addr add 192.0.2.1/24 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip -6 addr add 2001:DB8::1/32 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ grep 'release 6' /etc/redhat-release\n+ ip link set test1p master testbr\n+ ip link set test2p master testbr\n+ systemctl is-active firewalld\ninactive\n+ dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "start": "2024-09-24 14:50:29.565983", "end": "2024-09-24 14:50:30.776895", "delta": "0:00:01.210912", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203830.79433: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203830.79437: stdout chunk (state=3): >>><<< 13731 1727203830.79538: stderr chunk (state=3): >>><<< 13731 1727203830.79543: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ ip link add test1 type veth peer name test1p\n+ ip link add test2 type veth peer name test2p\n++ pgrep NetworkManager\n+ '[' -n 704 ']'\n+ nmcli d set test1 managed true\n+ nmcli d set test2 managed true\n+ nmcli d set test1p managed false\n+ nmcli d set test2p managed false\n+ ip link set test1p up\n+ ip link set test2p up\n+ ip link add name testbr type bridge forward_delay 0\n++ pgrep NetworkManager\n+ '[' -n 704 ']'\n+ nmcli d set testbr managed false\n+ ip link set testbr up\n+ timer=0\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ let timer+=1\n+ '[' 1 -eq 30 ']'\n+ sleep 1\n+ rc=0\n+ ip addr add 192.0.2.1/24 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip -6 addr add 2001:DB8::1/32 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ grep 'release 6' /etc/redhat-release\n+ ip link set test1p master testbr\n+ ip link set test2p master testbr\n+ systemctl is-active firewalld\ninactive\n+ dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "start": "2024-09-24 14:50:29.565983", "end": "2024-09-24 14:50:30.776895", "delta": "0:00:01.210912", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203830.79557: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n "$(pgrep NetworkManager)" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the \'testbr\' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n "$(pgrep NetworkManager)" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q \'inet [1-9]\'\ndo\n let "timer+=1"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc="$?"\n if [ "$rc" != 0 ]; then\n echo NOTICE - could not add testbr - error code "$rc"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc="$?"\n if [ "$rc" != 0 ]; then\n echo NOTICE - could not add testbr - error code "$rc"\n continue\n fi\ndone\n\nif grep \'release 6\' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo \'interface testbr {\' > /etc/radvd.conf\n echo \' AdvSendAdvert on;\' >> /etc/radvd.conf\n echo \' prefix 2001:DB8::/64 { \' >> /etc/radvd.conf\n echo \' AdvOnLink on; }; \' >> /etc/radvd.conf\n echo \' }; \' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service="$service"; then\n firewall-cmd --add-service "$service"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203830.79579: _low_level_execute_command(): starting 13731 1727203830.79590: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203829.2838345-14362-210838854669334/ > /dev/null 2>&1 && sleep 0' 13731 1727203830.80250: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203830.80291: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203830.80307: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203830.80393: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203830.80412: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203830.80465: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203830.80497: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203830.82459: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203830.82665: stderr chunk (state=3): >>><<< 13731 1727203830.82669: stdout chunk (state=3): >>><<< 13731 1727203830.82673: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203830.82679: handler run complete 13731 1727203830.82684: Evaluated conditional (False): False 13731 1727203830.82772: attempt loop complete, returning result 13731 1727203830.82777: _execute() done 13731 1727203830.82780: dumping result to json 13731 1727203830.82782: done dumping result, returning 13731 1727203830.82784: done running TaskExecutor() for managed-node3/TASK: Create test interfaces [028d2410-947f-82dc-c122-000000000115] 13731 1727203830.82786: sending task result for task 028d2410-947f-82dc-c122-000000000115 13731 1727203830.83185: done sending task result for task 028d2410-947f-82dc-c122-000000000115 13731 1727203830.83189: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "delta": "0:00:01.210912", "end": "2024-09-24 14:50:30.776895", "rc": 0, "start": "2024-09-24 14:50:29.565983" } STDERR: + exec + ip link add test1 type veth peer name test1p + ip link add test2 type veth peer name test2p ++ pgrep NetworkManager + '[' -n 704 ']' + nmcli d set test1 managed true + nmcli d set test2 managed true + nmcli d set test1p managed false + nmcli d set test2p managed false + ip link set test1p up + ip link set test2p up + ip link add name testbr type bridge forward_delay 0 ++ pgrep NetworkManager + '[' -n 704 ']' + nmcli d set testbr managed false + ip link set testbr up + timer=0 + ip addr show testbr + grep -q 'inet [1-9]' + let timer+=1 + '[' 1 -eq 30 ']' + sleep 1 + rc=0 + ip addr add 192.0.2.1/24 dev testbr + '[' 0 '!=' 0 ']' + ip -6 addr add 2001:DB8::1/32 dev testbr + '[' 0 '!=' 0 ']' + ip addr show testbr + grep -q 'inet [1-9]' + grep 'release 6' /etc/redhat-release + ip link set test1p master testbr + ip link set test2p master testbr + systemctl is-active firewalld inactive + dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces 13731 1727203830.83275: no more pending results, returning what we have 13731 1727203830.83281: results queue empty 13731 1727203830.83282: checking for any_errors_fatal 13731 1727203830.83289: done checking for any_errors_fatal 13731 1727203830.83290: checking for max_fail_percentage 13731 1727203830.83291: done checking for max_fail_percentage 13731 1727203830.83292: checking to see if all hosts have failed and the running result is not ok 13731 1727203830.83293: done checking to see if all hosts have failed 13731 1727203830.83293: getting the remaining hosts for this loop 13731 1727203830.83296: done getting the remaining hosts for this loop 13731 1727203830.83301: getting the next task for host managed-node3 13731 1727203830.83310: done getting next task for host managed-node3 13731 1727203830.83314: ^ task is: TASK: Include the task 'get_interface_stat.yml' 13731 1727203830.83318: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203830.83321: getting variables 13731 1727203830.83323: in VariableManager get_vars() 13731 1727203830.83352: Calling all_inventory to load vars for managed-node3 13731 1727203830.83355: Calling groups_inventory to load vars for managed-node3 13731 1727203830.83359: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203830.83373: Calling all_plugins_play to load vars for managed-node3 13731 1727203830.83582: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203830.83588: Calling groups_plugins_play to load vars for managed-node3 13731 1727203830.84171: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203830.84777: done with get_vars() 13731 1727203830.84790: done getting variables TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:3 Tuesday 24 September 2024 14:50:30 -0400 (0:00:01.671) 0:00:09.079 ***** 13731 1727203830.84888: entering _queue_task() for managed-node3/include_tasks 13731 1727203830.85191: worker is 1 (out of 1 available) 13731 1727203830.85203: exiting _queue_task() for managed-node3/include_tasks 13731 1727203830.85214: done queuing things up, now waiting for results queue to drain 13731 1727203830.85216: waiting for pending results... 13731 1727203830.85470: running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' 13731 1727203830.85601: in run() - task 028d2410-947f-82dc-c122-00000000011c 13731 1727203830.85624: variable 'ansible_search_path' from source: unknown 13731 1727203830.85679: variable 'ansible_search_path' from source: unknown 13731 1727203830.85683: calling self._execute() 13731 1727203830.85762: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203830.85777: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203830.85794: variable 'omit' from source: magic vars 13731 1727203830.86381: variable 'ansible_distribution_major_version' from source: facts 13731 1727203830.86385: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203830.86388: _execute() done 13731 1727203830.86391: dumping result to json 13731 1727203830.86393: done dumping result, returning 13731 1727203830.86396: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' [028d2410-947f-82dc-c122-00000000011c] 13731 1727203830.86397: sending task result for task 028d2410-947f-82dc-c122-00000000011c 13731 1727203830.86461: done sending task result for task 028d2410-947f-82dc-c122-00000000011c 13731 1727203830.86464: WORKER PROCESS EXITING 13731 1727203830.86488: no more pending results, returning what we have 13731 1727203830.86493: in VariableManager get_vars() 13731 1727203830.86522: Calling all_inventory to load vars for managed-node3 13731 1727203830.86524: Calling groups_inventory to load vars for managed-node3 13731 1727203830.86527: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203830.86538: Calling all_plugins_play to load vars for managed-node3 13731 1727203830.86541: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203830.86543: Calling groups_plugins_play to load vars for managed-node3 13731 1727203830.86805: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203830.87000: done with get_vars() 13731 1727203830.87008: variable 'ansible_search_path' from source: unknown 13731 1727203830.87009: variable 'ansible_search_path' from source: unknown 13731 1727203830.87044: we have included files to process 13731 1727203830.87046: generating all_blocks data 13731 1727203830.87047: done generating all_blocks data 13731 1727203830.87053: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203830.87055: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203830.87057: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203830.87287: done processing included file 13731 1727203830.87289: iterating over new_blocks loaded from include file 13731 1727203830.87291: in VariableManager get_vars() 13731 1727203830.87307: done with get_vars() 13731 1727203830.87308: filtering new block on tags 13731 1727203830.87337: done filtering new block on tags 13731 1727203830.87339: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node3 13731 1727203830.87344: extending task lists for all hosts with included blocks 13731 1727203830.87580: done extending task lists 13731 1727203830.87582: done processing included files 13731 1727203830.87582: results queue empty 13731 1727203830.87583: checking for any_errors_fatal 13731 1727203830.87588: done checking for any_errors_fatal 13731 1727203830.87589: checking for max_fail_percentage 13731 1727203830.87590: done checking for max_fail_percentage 13731 1727203830.87591: checking to see if all hosts have failed and the running result is not ok 13731 1727203830.87591: done checking to see if all hosts have failed 13731 1727203830.87592: getting the remaining hosts for this loop 13731 1727203830.87593: done getting the remaining hosts for this loop 13731 1727203830.87595: getting the next task for host managed-node3 13731 1727203830.87600: done getting next task for host managed-node3 13731 1727203830.87602: ^ task is: TASK: Get stat for interface {{ interface }} 13731 1727203830.87606: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203830.87608: getting variables 13731 1727203830.87609: in VariableManager get_vars() 13731 1727203830.87617: Calling all_inventory to load vars for managed-node3 13731 1727203830.87619: Calling groups_inventory to load vars for managed-node3 13731 1727203830.87621: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203830.87626: Calling all_plugins_play to load vars for managed-node3 13731 1727203830.87628: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203830.87631: Calling groups_plugins_play to load vars for managed-node3 13731 1727203830.87766: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203830.87960: done with get_vars() 13731 1727203830.87969: done getting variables 13731 1727203830.88124: variable 'interface' from source: task vars 13731 1727203830.88128: variable 'dhcp_interface1' from source: play vars 13731 1727203830.88188: variable 'dhcp_interface1' from source: play vars TASK [Get stat for interface test1] ******************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 Tuesday 24 September 2024 14:50:30 -0400 (0:00:00.033) 0:00:09.113 ***** 13731 1727203830.88219: entering _queue_task() for managed-node3/stat 13731 1727203830.88527: worker is 1 (out of 1 available) 13731 1727203830.88539: exiting _queue_task() for managed-node3/stat 13731 1727203830.88550: done queuing things up, now waiting for results queue to drain 13731 1727203830.88552: waiting for pending results... 13731 1727203830.88805: running TaskExecutor() for managed-node3/TASK: Get stat for interface test1 13731 1727203830.88948: in run() - task 028d2410-947f-82dc-c122-00000000017b 13731 1727203830.88969: variable 'ansible_search_path' from source: unknown 13731 1727203830.88979: variable 'ansible_search_path' from source: unknown 13731 1727203830.89020: calling self._execute() 13731 1727203830.89100: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203830.89116: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203830.89180: variable 'omit' from source: magic vars 13731 1727203830.89481: variable 'ansible_distribution_major_version' from source: facts 13731 1727203830.89498: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203830.89508: variable 'omit' from source: magic vars 13731 1727203830.89581: variable 'omit' from source: magic vars 13731 1727203830.89687: variable 'interface' from source: task vars 13731 1727203830.89696: variable 'dhcp_interface1' from source: play vars 13731 1727203830.89760: variable 'dhcp_interface1' from source: play vars 13731 1727203830.89790: variable 'omit' from source: magic vars 13731 1727203830.89980: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203830.89984: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203830.89986: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203830.89988: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203830.89990: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203830.89993: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203830.89995: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203830.89997: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203830.90061: Set connection var ansible_pipelining to False 13731 1727203830.90070: Set connection var ansible_shell_type to sh 13731 1727203830.90082: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203830.90090: Set connection var ansible_connection to ssh 13731 1727203830.90097: Set connection var ansible_shell_executable to /bin/sh 13731 1727203830.90104: Set connection var ansible_timeout to 10 13731 1727203830.90131: variable 'ansible_shell_executable' from source: unknown 13731 1727203830.90137: variable 'ansible_connection' from source: unknown 13731 1727203830.90142: variable 'ansible_module_compression' from source: unknown 13731 1727203830.90147: variable 'ansible_shell_type' from source: unknown 13731 1727203830.90152: variable 'ansible_shell_executable' from source: unknown 13731 1727203830.90156: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203830.90162: variable 'ansible_pipelining' from source: unknown 13731 1727203830.90167: variable 'ansible_timeout' from source: unknown 13731 1727203830.90173: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203830.90373: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203830.90391: variable 'omit' from source: magic vars 13731 1727203830.90399: starting attempt loop 13731 1727203830.90404: running the handler 13731 1727203830.90420: _low_level_execute_command(): starting 13731 1727203830.90563: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203830.91164: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203830.91195: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203830.91234: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203830.91319: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203830.91335: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203830.91353: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203830.91421: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203830.93001: stdout chunk (state=3): >>>/root <<< 13731 1727203830.93168: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203830.93172: stdout chunk (state=3): >>><<< 13731 1727203830.93174: stderr chunk (state=3): >>><<< 13731 1727203830.93202: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203830.93322: _low_level_execute_command(): starting 13731 1727203830.93328: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614 `" && echo ansible-tmp-1727203830.932292-14432-159427700955614="` echo /root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614 `" ) && sleep 0' 13731 1727203830.94491: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203830.94501: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203830.94514: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203830.94715: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203830.94736: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203830.94751: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203830.96852: stdout chunk (state=3): >>>ansible-tmp-1727203830.932292-14432-159427700955614=/root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614 <<< 13731 1727203830.96901: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203830.96950: stderr chunk (state=3): >>><<< 13731 1727203830.96953: stdout chunk (state=3): >>><<< 13731 1727203830.96971: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203830.932292-14432-159427700955614=/root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203830.97380: variable 'ansible_module_compression' from source: unknown 13731 1727203830.97384: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13731 1727203830.97386: variable 'ansible_facts' from source: unknown 13731 1727203830.97433: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/AnsiballZ_stat.py 13731 1727203830.97760: Sending initial data 13731 1727203830.97806: Sent initial data (152 bytes) 13731 1727203830.98411: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203830.98424: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203830.98457: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203830.98505: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203830.98566: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203830.98583: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.00099: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203831.00209: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203831.00244: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/AnsiballZ_stat.py" <<< 13731 1727203831.00308: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpmzuytc06 /root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/AnsiballZ_stat.py <<< 13731 1727203831.00324: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpmzuytc06" to remote "/root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/AnsiballZ_stat.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/AnsiballZ_stat.py" <<< 13731 1727203831.01451: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203831.01462: stdout chunk (state=3): >>><<< 13731 1727203831.01477: stderr chunk (state=3): >>><<< 13731 1727203831.01688: done transferring module to remote 13731 1727203831.01691: _low_level_execute_command(): starting 13731 1727203831.01693: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/ /root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/AnsiballZ_stat.py && sleep 0' 13731 1727203831.02369: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203831.02389: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203831.02407: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203831.02426: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203831.02530: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203831.02552: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203831.02568: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.02638: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.04406: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203831.04410: stdout chunk (state=3): >>><<< 13731 1727203831.04412: stderr chunk (state=3): >>><<< 13731 1727203831.04449: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203831.04467: _low_level_execute_command(): starting 13731 1727203831.04478: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/AnsiballZ_stat.py && sleep 0' 13731 1727203831.05592: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203831.05609: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203831.05624: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203831.05645: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203831.05747: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203831.05772: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.05844: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.20793: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test1", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 27505, "dev": 23, "nlink": 1, "atime": 1727203829.5722764, "mtime": 1727203829.5722764, "ctime": 1727203829.5722764, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test1", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13731 1727203831.22107: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203831.22111: stdout chunk (state=3): >>><<< 13731 1727203831.22116: stderr chunk (state=3): >>><<< 13731 1727203831.22281: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test1", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 27505, "dev": 23, "nlink": 1, "atime": 1727203829.5722764, "mtime": 1727203829.5722764, "ctime": 1727203829.5722764, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test1", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203831.22285: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/sys/class/net/test1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203831.22287: _low_level_execute_command(): starting 13731 1727203831.22289: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203830.932292-14432-159427700955614/ > /dev/null 2>&1 && sleep 0' 13731 1727203831.23227: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203831.23246: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203831.23294: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203831.23307: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203831.23321: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203831.23395: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.23460: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.23492: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.25357: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203831.25386: stdout chunk (state=3): >>><<< 13731 1727203831.25401: stderr chunk (state=3): >>><<< 13731 1727203831.25422: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203831.25434: handler run complete 13731 1727203831.25502: attempt loop complete, returning result 13731 1727203831.25509: _execute() done 13731 1727203831.25515: dumping result to json 13731 1727203831.25580: done dumping result, returning 13731 1727203831.25585: done running TaskExecutor() for managed-node3/TASK: Get stat for interface test1 [028d2410-947f-82dc-c122-00000000017b] 13731 1727203831.25588: sending task result for task 028d2410-947f-82dc-c122-00000000017b ok: [managed-node3] => { "changed": false, "stat": { "atime": 1727203829.5722764, "block_size": 4096, "blocks": 0, "ctime": 1727203829.5722764, "dev": 23, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 27505, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": true, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "mode": "0777", "mtime": 1727203829.5722764, "nlink": 1, "path": "/sys/class/net/test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "wgrp": true, "woth": true, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 13731 1727203831.25870: no more pending results, returning what we have 13731 1727203831.25874: results queue empty 13731 1727203831.25876: checking for any_errors_fatal 13731 1727203831.25878: done checking for any_errors_fatal 13731 1727203831.25879: checking for max_fail_percentage 13731 1727203831.25880: done checking for max_fail_percentage 13731 1727203831.25881: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.25882: done checking to see if all hosts have failed 13731 1727203831.25882: getting the remaining hosts for this loop 13731 1727203831.25884: done getting the remaining hosts for this loop 13731 1727203831.25889: getting the next task for host managed-node3 13731 1727203831.25897: done getting next task for host managed-node3 13731 1727203831.25900: ^ task is: TASK: Assert that the interface is present - '{{ interface }}' 13731 1727203831.25904: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.25908: getting variables 13731 1727203831.25909: in VariableManager get_vars() 13731 1727203831.25939: Calling all_inventory to load vars for managed-node3 13731 1727203831.25942: Calling groups_inventory to load vars for managed-node3 13731 1727203831.25945: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.25956: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.25959: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.25964: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.27012: done sending task result for task 028d2410-947f-82dc-c122-00000000017b 13731 1727203831.27016: WORKER PROCESS EXITING 13731 1727203831.27206: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.27691: done with get_vars() 13731 1727203831.27703: done getting variables 13731 1727203831.27797: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) 13731 1727203831.28119: variable 'interface' from source: task vars 13731 1727203831.28123: variable 'dhcp_interface1' from source: play vars 13731 1727203831.28195: variable 'dhcp_interface1' from source: play vars TASK [Assert that the interface is present - 'test1'] ************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:5 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.400) 0:00:09.513 ***** 13731 1727203831.28232: entering _queue_task() for managed-node3/assert 13731 1727203831.28234: Creating lock for assert 13731 1727203831.28562: worker is 1 (out of 1 available) 13731 1727203831.28574: exiting _queue_task() for managed-node3/assert 13731 1727203831.28588: done queuing things up, now waiting for results queue to drain 13731 1727203831.28589: waiting for pending results... 13731 1727203831.28845: running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'test1' 13731 1727203831.28991: in run() - task 028d2410-947f-82dc-c122-00000000011d 13731 1727203831.29010: variable 'ansible_search_path' from source: unknown 13731 1727203831.29018: variable 'ansible_search_path' from source: unknown 13731 1727203831.29059: calling self._execute() 13731 1727203831.29148: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.29176: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.29195: variable 'omit' from source: magic vars 13731 1727203831.29566: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.29587: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.29599: variable 'omit' from source: magic vars 13731 1727203831.29672: variable 'omit' from source: magic vars 13731 1727203831.29784: variable 'interface' from source: task vars 13731 1727203831.29794: variable 'dhcp_interface1' from source: play vars 13731 1727203831.29866: variable 'dhcp_interface1' from source: play vars 13731 1727203831.29893: variable 'omit' from source: magic vars 13731 1727203831.29940: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203831.30056: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203831.30059: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203831.30061: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203831.30063: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203831.30071: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203831.30082: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.30091: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.30205: Set connection var ansible_pipelining to False 13731 1727203831.30215: Set connection var ansible_shell_type to sh 13731 1727203831.30225: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203831.30233: Set connection var ansible_connection to ssh 13731 1727203831.30243: Set connection var ansible_shell_executable to /bin/sh 13731 1727203831.30252: Set connection var ansible_timeout to 10 13731 1727203831.30284: variable 'ansible_shell_executable' from source: unknown 13731 1727203831.30291: variable 'ansible_connection' from source: unknown 13731 1727203831.30298: variable 'ansible_module_compression' from source: unknown 13731 1727203831.30305: variable 'ansible_shell_type' from source: unknown 13731 1727203831.30312: variable 'ansible_shell_executable' from source: unknown 13731 1727203831.30318: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.30326: variable 'ansible_pipelining' from source: unknown 13731 1727203831.30382: variable 'ansible_timeout' from source: unknown 13731 1727203831.30385: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.30486: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203831.30506: variable 'omit' from source: magic vars 13731 1727203831.30517: starting attempt loop 13731 1727203831.30525: running the handler 13731 1727203831.30661: variable 'interface_stat' from source: set_fact 13731 1727203831.30690: Evaluated conditional (interface_stat.stat.exists): True 13731 1727203831.30700: handler run complete 13731 1727203831.30780: attempt loop complete, returning result 13731 1727203831.30783: _execute() done 13731 1727203831.30786: dumping result to json 13731 1727203831.30788: done dumping result, returning 13731 1727203831.30790: done running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'test1' [028d2410-947f-82dc-c122-00000000011d] 13731 1727203831.30792: sending task result for task 028d2410-947f-82dc-c122-00000000011d 13731 1727203831.31081: done sending task result for task 028d2410-947f-82dc-c122-00000000011d 13731 1727203831.31085: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203831.31126: no more pending results, returning what we have 13731 1727203831.31130: results queue empty 13731 1727203831.31131: checking for any_errors_fatal 13731 1727203831.31139: done checking for any_errors_fatal 13731 1727203831.31140: checking for max_fail_percentage 13731 1727203831.31142: done checking for max_fail_percentage 13731 1727203831.31143: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.31143: done checking to see if all hosts have failed 13731 1727203831.31144: getting the remaining hosts for this loop 13731 1727203831.31146: done getting the remaining hosts for this loop 13731 1727203831.31149: getting the next task for host managed-node3 13731 1727203831.31158: done getting next task for host managed-node3 13731 1727203831.31161: ^ task is: TASK: Include the task 'get_interface_stat.yml' 13731 1727203831.31166: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.31170: getting variables 13731 1727203831.31171: in VariableManager get_vars() 13731 1727203831.31201: Calling all_inventory to load vars for managed-node3 13731 1727203831.31204: Calling groups_inventory to load vars for managed-node3 13731 1727203831.31207: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.31215: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.31217: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.31219: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.31456: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.31644: done with get_vars() 13731 1727203831.31656: done getting variables TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:3 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.035) 0:00:09.548 ***** 13731 1727203831.31758: entering _queue_task() for managed-node3/include_tasks 13731 1727203831.32037: worker is 1 (out of 1 available) 13731 1727203831.32050: exiting _queue_task() for managed-node3/include_tasks 13731 1727203831.32062: done queuing things up, now waiting for results queue to drain 13731 1727203831.32064: waiting for pending results... 13731 1727203831.32318: running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' 13731 1727203831.32444: in run() - task 028d2410-947f-82dc-c122-000000000121 13731 1727203831.32466: variable 'ansible_search_path' from source: unknown 13731 1727203831.32478: variable 'ansible_search_path' from source: unknown 13731 1727203831.32521: calling self._execute() 13731 1727203831.32609: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.32620: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.32633: variable 'omit' from source: magic vars 13731 1727203831.33003: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.33020: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.33029: _execute() done 13731 1727203831.33038: dumping result to json 13731 1727203831.33047: done dumping result, returning 13731 1727203831.33057: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' [028d2410-947f-82dc-c122-000000000121] 13731 1727203831.33067: sending task result for task 028d2410-947f-82dc-c122-000000000121 13731 1727203831.33383: done sending task result for task 028d2410-947f-82dc-c122-000000000121 13731 1727203831.33386: WORKER PROCESS EXITING 13731 1727203831.33411: no more pending results, returning what we have 13731 1727203831.33415: in VariableManager get_vars() 13731 1727203831.33445: Calling all_inventory to load vars for managed-node3 13731 1727203831.33448: Calling groups_inventory to load vars for managed-node3 13731 1727203831.33451: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.33462: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.33465: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.33468: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.33760: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.33945: done with get_vars() 13731 1727203831.33953: variable 'ansible_search_path' from source: unknown 13731 1727203831.33954: variable 'ansible_search_path' from source: unknown 13731 1727203831.33992: we have included files to process 13731 1727203831.33993: generating all_blocks data 13731 1727203831.33995: done generating all_blocks data 13731 1727203831.34000: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203831.34001: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203831.34003: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203831.34181: done processing included file 13731 1727203831.34184: iterating over new_blocks loaded from include file 13731 1727203831.34185: in VariableManager get_vars() 13731 1727203831.34202: done with get_vars() 13731 1727203831.34203: filtering new block on tags 13731 1727203831.34232: done filtering new block on tags 13731 1727203831.34235: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node3 13731 1727203831.34240: extending task lists for all hosts with included blocks 13731 1727203831.34440: done extending task lists 13731 1727203831.34442: done processing included files 13731 1727203831.34442: results queue empty 13731 1727203831.34443: checking for any_errors_fatal 13731 1727203831.34446: done checking for any_errors_fatal 13731 1727203831.34447: checking for max_fail_percentage 13731 1727203831.34448: done checking for max_fail_percentage 13731 1727203831.34449: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.34450: done checking to see if all hosts have failed 13731 1727203831.34450: getting the remaining hosts for this loop 13731 1727203831.34451: done getting the remaining hosts for this loop 13731 1727203831.34454: getting the next task for host managed-node3 13731 1727203831.34458: done getting next task for host managed-node3 13731 1727203831.34461: ^ task is: TASK: Get stat for interface {{ interface }} 13731 1727203831.34465: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.34467: getting variables 13731 1727203831.34468: in VariableManager get_vars() 13731 1727203831.34478: Calling all_inventory to load vars for managed-node3 13731 1727203831.34481: Calling groups_inventory to load vars for managed-node3 13731 1727203831.34483: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.34489: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.34491: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.34494: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.34634: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.34825: done with get_vars() 13731 1727203831.34834: done getting variables 13731 1727203831.34988: variable 'interface' from source: task vars 13731 1727203831.34992: variable 'dhcp_interface2' from source: play vars 13731 1727203831.35052: variable 'dhcp_interface2' from source: play vars TASK [Get stat for interface test2] ******************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.034) 0:00:09.582 ***** 13731 1727203831.35190: entering _queue_task() for managed-node3/stat 13731 1727203831.36013: worker is 1 (out of 1 available) 13731 1727203831.36026: exiting _queue_task() for managed-node3/stat 13731 1727203831.36038: done queuing things up, now waiting for results queue to drain 13731 1727203831.36040: waiting for pending results... 13731 1727203831.36669: running TaskExecutor() for managed-node3/TASK: Get stat for interface test2 13731 1727203831.37144: in run() - task 028d2410-947f-82dc-c122-00000000019f 13731 1727203831.37171: variable 'ansible_search_path' from source: unknown 13731 1727203831.37206: variable 'ansible_search_path' from source: unknown 13731 1727203831.37481: calling self._execute() 13731 1727203831.37745: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.37750: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.37753: variable 'omit' from source: magic vars 13731 1727203831.39047: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.39493: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.39499: variable 'omit' from source: magic vars 13731 1727203831.39681: variable 'omit' from source: magic vars 13731 1727203831.40039: variable 'interface' from source: task vars 13731 1727203831.40043: variable 'dhcp_interface2' from source: play vars 13731 1727203831.40160: variable 'dhcp_interface2' from source: play vars 13731 1727203831.40230: variable 'omit' from source: magic vars 13731 1727203831.40396: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203831.40436: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203831.40497: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203831.40521: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203831.40681: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203831.40689: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203831.40692: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.40694: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.40857: Set connection var ansible_pipelining to False 13731 1727203831.41024: Set connection var ansible_shell_type to sh 13731 1727203831.41035: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203831.41045: Set connection var ansible_connection to ssh 13731 1727203831.41055: Set connection var ansible_shell_executable to /bin/sh 13731 1727203831.41281: Set connection var ansible_timeout to 10 13731 1727203831.41284: variable 'ansible_shell_executable' from source: unknown 13731 1727203831.41286: variable 'ansible_connection' from source: unknown 13731 1727203831.41288: variable 'ansible_module_compression' from source: unknown 13731 1727203831.41290: variable 'ansible_shell_type' from source: unknown 13731 1727203831.41291: variable 'ansible_shell_executable' from source: unknown 13731 1727203831.41293: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.41295: variable 'ansible_pipelining' from source: unknown 13731 1727203831.41298: variable 'ansible_timeout' from source: unknown 13731 1727203831.41299: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.41649: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203831.41670: variable 'omit' from source: magic vars 13731 1727203831.41869: starting attempt loop 13731 1727203831.41873: running the handler 13731 1727203831.41878: _low_level_execute_command(): starting 13731 1727203831.41882: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203831.43616: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.43666: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203831.43672: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.43715: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.45370: stdout chunk (state=3): >>>/root <<< 13731 1727203831.45516: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203831.45520: stdout chunk (state=3): >>><<< 13731 1727203831.45522: stderr chunk (state=3): >>><<< 13731 1727203831.45672: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203831.45677: _low_level_execute_command(): starting 13731 1727203831.45681: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212 `" && echo ansible-tmp-1727203831.455529-14455-153890835005212="` echo /root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212 `" ) && sleep 0' 13731 1727203831.46840: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203831.46942: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.47087: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203831.47179: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.47219: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.49077: stdout chunk (state=3): >>>ansible-tmp-1727203831.455529-14455-153890835005212=/root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212 <<< 13731 1727203831.49352: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203831.49356: stdout chunk (state=3): >>><<< 13731 1727203831.49358: stderr chunk (state=3): >>><<< 13731 1727203831.49464: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203831.455529-14455-153890835005212=/root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203831.49468: variable 'ansible_module_compression' from source: unknown 13731 1727203831.49522: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13731 1727203831.49573: variable 'ansible_facts' from source: unknown 13731 1727203831.49696: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/AnsiballZ_stat.py 13731 1727203831.49920: Sending initial data 13731 1727203831.49923: Sent initial data (152 bytes) 13731 1727203831.50495: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203831.50509: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203831.50522: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203831.50538: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203831.50565: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203831.50591: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.50678: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.50702: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203831.50721: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203831.50746: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.50850: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.52418: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203831.52485: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/AnsiballZ_stat.py" <<< 13731 1727203831.52489: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpdlfpm82x /root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/AnsiballZ_stat.py <<< 13731 1727203831.52619: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpdlfpm82x" to remote "/root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/AnsiballZ_stat.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/AnsiballZ_stat.py" <<< 13731 1727203831.53374: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203831.53417: stderr chunk (state=3): >>><<< 13731 1727203831.53420: stdout chunk (state=3): >>><<< 13731 1727203831.53429: done transferring module to remote 13731 1727203831.53442: _low_level_execute_command(): starting 13731 1727203831.53449: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/ /root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/AnsiballZ_stat.py && sleep 0' 13731 1727203831.54130: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203831.54190: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203831.54204: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203831.54270: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.54344: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203831.54382: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203831.54440: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.54469: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.56253: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203831.56284: stdout chunk (state=3): >>><<< 13731 1727203831.56287: stderr chunk (state=3): >>><<< 13731 1727203831.56382: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203831.56385: _low_level_execute_command(): starting 13731 1727203831.56387: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/AnsiballZ_stat.py && sleep 0' 13731 1727203831.56949: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203831.56961: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203831.56989: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203831.57043: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.57105: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203831.57121: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203831.57151: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.57220: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.72424: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test2", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 27911, "dev": 23, "nlink": 1, "atime": 1727203829.5779638, "mtime": 1727203829.5779638, "ctime": 1727203829.5779638, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test2", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13731 1727203831.73728: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203831.73750: stderr chunk (state=3): >>><<< 13731 1727203831.73753: stdout chunk (state=3): >>><<< 13731 1727203831.73777: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test2", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 27911, "dev": 23, "nlink": 1, "atime": 1727203829.5779638, "mtime": 1727203829.5779638, "ctime": 1727203829.5779638, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test2", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203831.73814: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/sys/class/net/test2', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203831.73822: _low_level_execute_command(): starting 13731 1727203831.73826: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203831.455529-14455-153890835005212/ > /dev/null 2>&1 && sleep 0' 13731 1727203831.74314: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203831.74321: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203831.74324: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.74341: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203831.74401: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203831.74420: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203831.74466: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203831.76260: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203831.76287: stderr chunk (state=3): >>><<< 13731 1727203831.76290: stdout chunk (state=3): >>><<< 13731 1727203831.76303: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203831.76312: handler run complete 13731 1727203831.76342: attempt loop complete, returning result 13731 1727203831.76345: _execute() done 13731 1727203831.76348: dumping result to json 13731 1727203831.76351: done dumping result, returning 13731 1727203831.76360: done running TaskExecutor() for managed-node3/TASK: Get stat for interface test2 [028d2410-947f-82dc-c122-00000000019f] 13731 1727203831.76365: sending task result for task 028d2410-947f-82dc-c122-00000000019f 13731 1727203831.76471: done sending task result for task 028d2410-947f-82dc-c122-00000000019f 13731 1727203831.76474: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "stat": { "atime": 1727203829.5779638, "block_size": 4096, "blocks": 0, "ctime": 1727203829.5779638, "dev": 23, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 27911, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": true, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "mode": "0777", "mtime": 1727203829.5779638, "nlink": 1, "path": "/sys/class/net/test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "wgrp": true, "woth": true, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 13731 1727203831.76639: no more pending results, returning what we have 13731 1727203831.76643: results queue empty 13731 1727203831.76643: checking for any_errors_fatal 13731 1727203831.76645: done checking for any_errors_fatal 13731 1727203831.76645: checking for max_fail_percentage 13731 1727203831.76646: done checking for max_fail_percentage 13731 1727203831.76647: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.76648: done checking to see if all hosts have failed 13731 1727203831.76648: getting the remaining hosts for this loop 13731 1727203831.76650: done getting the remaining hosts for this loop 13731 1727203831.76653: getting the next task for host managed-node3 13731 1727203831.76660: done getting next task for host managed-node3 13731 1727203831.76665: ^ task is: TASK: Assert that the interface is present - '{{ interface }}' 13731 1727203831.76668: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.76672: getting variables 13731 1727203831.76673: in VariableManager get_vars() 13731 1727203831.76700: Calling all_inventory to load vars for managed-node3 13731 1727203831.76702: Calling groups_inventory to load vars for managed-node3 13731 1727203831.76704: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.76711: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.76712: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.76714: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.76816: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.76934: done with get_vars() 13731 1727203831.76941: done getting variables 13731 1727203831.76990: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203831.77082: variable 'interface' from source: task vars 13731 1727203831.77085: variable 'dhcp_interface2' from source: play vars 13731 1727203831.77128: variable 'dhcp_interface2' from source: play vars TASK [Assert that the interface is present - 'test2'] ************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:5 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.419) 0:00:10.002 ***** 13731 1727203831.77154: entering _queue_task() for managed-node3/assert 13731 1727203831.77351: worker is 1 (out of 1 available) 13731 1727203831.77365: exiting _queue_task() for managed-node3/assert 13731 1727203831.77379: done queuing things up, now waiting for results queue to drain 13731 1727203831.77381: waiting for pending results... 13731 1727203831.77711: running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'test2' 13731 1727203831.77717: in run() - task 028d2410-947f-82dc-c122-000000000122 13731 1727203831.77720: variable 'ansible_search_path' from source: unknown 13731 1727203831.77722: variable 'ansible_search_path' from source: unknown 13731 1727203831.77764: calling self._execute() 13731 1727203831.77849: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.77865: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.77882: variable 'omit' from source: magic vars 13731 1727203831.78226: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.78297: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.78300: variable 'omit' from source: magic vars 13731 1727203831.78349: variable 'omit' from source: magic vars 13731 1727203831.78496: variable 'interface' from source: task vars 13731 1727203831.78500: variable 'dhcp_interface2' from source: play vars 13731 1727203831.78533: variable 'dhcp_interface2' from source: play vars 13731 1727203831.78560: variable 'omit' from source: magic vars 13731 1727203831.78595: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203831.78620: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203831.78635: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203831.78650: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203831.78661: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203831.78688: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203831.78691: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.78694: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.78770: Set connection var ansible_pipelining to False 13731 1727203831.78773: Set connection var ansible_shell_type to sh 13731 1727203831.78779: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203831.78784: Set connection var ansible_connection to ssh 13731 1727203831.78789: Set connection var ansible_shell_executable to /bin/sh 13731 1727203831.78794: Set connection var ansible_timeout to 10 13731 1727203831.78811: variable 'ansible_shell_executable' from source: unknown 13731 1727203831.78814: variable 'ansible_connection' from source: unknown 13731 1727203831.78816: variable 'ansible_module_compression' from source: unknown 13731 1727203831.78818: variable 'ansible_shell_type' from source: unknown 13731 1727203831.78821: variable 'ansible_shell_executable' from source: unknown 13731 1727203831.78823: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.78828: variable 'ansible_pipelining' from source: unknown 13731 1727203831.78830: variable 'ansible_timeout' from source: unknown 13731 1727203831.78833: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.78934: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203831.78942: variable 'omit' from source: magic vars 13731 1727203831.78947: starting attempt loop 13731 1727203831.78950: running the handler 13731 1727203831.79041: variable 'interface_stat' from source: set_fact 13731 1727203831.79057: Evaluated conditional (interface_stat.stat.exists): True 13731 1727203831.79062: handler run complete 13731 1727203831.79074: attempt loop complete, returning result 13731 1727203831.79078: _execute() done 13731 1727203831.79081: dumping result to json 13731 1727203831.79084: done dumping result, returning 13731 1727203831.79095: done running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'test2' [028d2410-947f-82dc-c122-000000000122] 13731 1727203831.79097: sending task result for task 028d2410-947f-82dc-c122-000000000122 13731 1727203831.79170: done sending task result for task 028d2410-947f-82dc-c122-000000000122 13731 1727203831.79173: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203831.79250: no more pending results, returning what we have 13731 1727203831.79253: results queue empty 13731 1727203831.79254: checking for any_errors_fatal 13731 1727203831.79261: done checking for any_errors_fatal 13731 1727203831.79262: checking for max_fail_percentage 13731 1727203831.79263: done checking for max_fail_percentage 13731 1727203831.79264: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.79264: done checking to see if all hosts have failed 13731 1727203831.79265: getting the remaining hosts for this loop 13731 1727203831.79267: done getting the remaining hosts for this loop 13731 1727203831.79270: getting the next task for host managed-node3 13731 1727203831.79279: done getting next task for host managed-node3 13731 1727203831.79282: ^ task is: TASK: Test 13731 1727203831.79284: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.79287: getting variables 13731 1727203831.79288: in VariableManager get_vars() 13731 1727203831.79318: Calling all_inventory to load vars for managed-node3 13731 1727203831.79321: Calling groups_inventory to load vars for managed-node3 13731 1727203831.79324: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.79332: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.79334: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.79336: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.79452: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.79593: done with get_vars() 13731 1727203831.79600: done getting variables TASK [Test] ******************************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:30 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.025) 0:00:10.027 ***** 13731 1727203831.79662: entering _queue_task() for managed-node3/include_tasks 13731 1727203831.79848: worker is 1 (out of 1 available) 13731 1727203831.79863: exiting _queue_task() for managed-node3/include_tasks 13731 1727203831.79874: done queuing things up, now waiting for results queue to drain 13731 1727203831.79878: waiting for pending results... 13731 1727203831.80024: running TaskExecutor() for managed-node3/TASK: Test 13731 1727203831.80084: in run() - task 028d2410-947f-82dc-c122-00000000008c 13731 1727203831.80096: variable 'ansible_search_path' from source: unknown 13731 1727203831.80102: variable 'ansible_search_path' from source: unknown 13731 1727203831.80135: variable 'lsr_test' from source: include params 13731 1727203831.80285: variable 'lsr_test' from source: include params 13731 1727203831.80335: variable 'omit' from source: magic vars 13731 1727203831.80421: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.80428: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.80437: variable 'omit' from source: magic vars 13731 1727203831.80601: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.80608: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.80613: variable 'item' from source: unknown 13731 1727203831.80658: variable 'item' from source: unknown 13731 1727203831.80686: variable 'item' from source: unknown 13731 1727203831.80727: variable 'item' from source: unknown 13731 1727203831.80835: dumping result to json 13731 1727203831.80838: done dumping result, returning 13731 1727203831.80840: done running TaskExecutor() for managed-node3/TASK: Test [028d2410-947f-82dc-c122-00000000008c] 13731 1727203831.80842: sending task result for task 028d2410-947f-82dc-c122-00000000008c 13731 1727203831.80880: done sending task result for task 028d2410-947f-82dc-c122-00000000008c 13731 1727203831.80901: no more pending results, returning what we have 13731 1727203831.80905: in VariableManager get_vars() 13731 1727203831.80933: Calling all_inventory to load vars for managed-node3 13731 1727203831.80935: Calling groups_inventory to load vars for managed-node3 13731 1727203831.80937: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.80947: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.80949: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.80951: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.81078: WORKER PROCESS EXITING 13731 1727203831.81089: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.81199: done with get_vars() 13731 1727203831.81206: variable 'ansible_search_path' from source: unknown 13731 1727203831.81207: variable 'ansible_search_path' from source: unknown 13731 1727203831.81232: we have included files to process 13731 1727203831.81233: generating all_blocks data 13731 1727203831.81234: done generating all_blocks data 13731 1727203831.81237: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile.yml 13731 1727203831.81238: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile.yml 13731 1727203831.81239: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile.yml 13731 1727203831.81521: done processing included file 13731 1727203831.81523: iterating over new_blocks loaded from include file 13731 1727203831.81524: in VariableManager get_vars() 13731 1727203831.81534: done with get_vars() 13731 1727203831.81536: filtering new block on tags 13731 1727203831.81556: done filtering new block on tags 13731 1727203831.81557: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile.yml for managed-node3 => (item=tasks/create_bond_profile.yml) 13731 1727203831.81560: extending task lists for all hosts with included blocks 13731 1727203831.82279: done extending task lists 13731 1727203831.82280: done processing included files 13731 1727203831.82281: results queue empty 13731 1727203831.82281: checking for any_errors_fatal 13731 1727203831.82283: done checking for any_errors_fatal 13731 1727203831.82284: checking for max_fail_percentage 13731 1727203831.82285: done checking for max_fail_percentage 13731 1727203831.82286: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.82286: done checking to see if all hosts have failed 13731 1727203831.82287: getting the remaining hosts for this loop 13731 1727203831.82288: done getting the remaining hosts for this loop 13731 1727203831.82290: getting the next task for host managed-node3 13731 1727203831.82293: done getting next task for host managed-node3 13731 1727203831.82295: ^ task is: TASK: Include network role 13731 1727203831.82297: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.82298: getting variables 13731 1727203831.82299: in VariableManager get_vars() 13731 1727203831.82305: Calling all_inventory to load vars for managed-node3 13731 1727203831.82306: Calling groups_inventory to load vars for managed-node3 13731 1727203831.82307: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.82311: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.82313: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.82314: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.82400: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.82508: done with get_vars() 13731 1727203831.82515: done getting variables TASK [Include network role] **************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile.yml:3 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.028) 0:00:10.056 ***** 13731 1727203831.82565: entering _queue_task() for managed-node3/include_role 13731 1727203831.82566: Creating lock for include_role 13731 1727203831.82791: worker is 1 (out of 1 available) 13731 1727203831.82802: exiting _queue_task() for managed-node3/include_role 13731 1727203831.82813: done queuing things up, now waiting for results queue to drain 13731 1727203831.82815: waiting for pending results... 13731 1727203831.82956: running TaskExecutor() for managed-node3/TASK: Include network role 13731 1727203831.83033: in run() - task 028d2410-947f-82dc-c122-0000000001c5 13731 1727203831.83045: variable 'ansible_search_path' from source: unknown 13731 1727203831.83049: variable 'ansible_search_path' from source: unknown 13731 1727203831.83083: calling self._execute() 13731 1727203831.83140: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.83145: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.83153: variable 'omit' from source: magic vars 13731 1727203831.83418: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.83428: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.83433: _execute() done 13731 1727203831.83436: dumping result to json 13731 1727203831.83438: done dumping result, returning 13731 1727203831.83444: done running TaskExecutor() for managed-node3/TASK: Include network role [028d2410-947f-82dc-c122-0000000001c5] 13731 1727203831.83449: sending task result for task 028d2410-947f-82dc-c122-0000000001c5 13731 1727203831.83583: done sending task result for task 028d2410-947f-82dc-c122-0000000001c5 13731 1727203831.83586: WORKER PROCESS EXITING 13731 1727203831.83612: no more pending results, returning what we have 13731 1727203831.83618: in VariableManager get_vars() 13731 1727203831.83648: Calling all_inventory to load vars for managed-node3 13731 1727203831.83650: Calling groups_inventory to load vars for managed-node3 13731 1727203831.83653: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.83663: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.83665: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.83667: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.83816: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.83934: done with get_vars() 13731 1727203831.83943: variable 'ansible_search_path' from source: unknown 13731 1727203831.83944: variable 'ansible_search_path' from source: unknown 13731 1727203831.84097: variable 'omit' from source: magic vars 13731 1727203831.84126: variable 'omit' from source: magic vars 13731 1727203831.84136: variable 'omit' from source: magic vars 13731 1727203831.84138: we have included files to process 13731 1727203831.84139: generating all_blocks data 13731 1727203831.84140: done generating all_blocks data 13731 1727203831.84141: processing included file: fedora.linux_system_roles.network 13731 1727203831.84155: in VariableManager get_vars() 13731 1727203831.84164: done with get_vars() 13731 1727203831.84211: in VariableManager get_vars() 13731 1727203831.84224: done with get_vars() 13731 1727203831.84258: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/defaults/main.yml 13731 1727203831.84415: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/meta/main.yml 13731 1727203831.84499: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml 13731 1727203831.84884: in VariableManager get_vars() 13731 1727203831.84897: done with get_vars() 13731 1727203831.85180: trying /usr/local/lib/python3.12/site-packages/ansible/modules/__pycache__ redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13731 1727203831.86366: iterating over new_blocks loaded from include file 13731 1727203831.86369: in VariableManager get_vars() 13731 1727203831.86388: done with get_vars() 13731 1727203831.86390: filtering new block on tags 13731 1727203831.86664: done filtering new block on tags 13731 1727203831.86668: in VariableManager get_vars() 13731 1727203831.86685: done with get_vars() 13731 1727203831.86687: filtering new block on tags 13731 1727203831.86704: done filtering new block on tags 13731 1727203831.86707: done iterating over new_blocks loaded from include file included: fedora.linux_system_roles.network for managed-node3 13731 1727203831.86713: extending task lists for all hosts with included blocks 13731 1727203831.86867: done extending task lists 13731 1727203831.86869: done processing included files 13731 1727203831.86870: results queue empty 13731 1727203831.86870: checking for any_errors_fatal 13731 1727203831.86874: done checking for any_errors_fatal 13731 1727203831.86877: checking for max_fail_percentage 13731 1727203831.86878: done checking for max_fail_percentage 13731 1727203831.86879: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.86880: done checking to see if all hosts have failed 13731 1727203831.86880: getting the remaining hosts for this loop 13731 1727203831.86882: done getting the remaining hosts for this loop 13731 1727203831.86884: getting the next task for host managed-node3 13731 1727203831.86888: done getting next task for host managed-node3 13731 1727203831.86890: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13731 1727203831.86893: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.86902: getting variables 13731 1727203831.86903: in VariableManager get_vars() 13731 1727203831.86915: Calling all_inventory to load vars for managed-node3 13731 1727203831.86917: Calling groups_inventory to load vars for managed-node3 13731 1727203831.86919: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.86925: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.86927: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.86930: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.87245: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.87427: done with get_vars() 13731 1727203831.87433: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:4 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.049) 0:00:10.106 ***** 13731 1727203831.87508: entering _queue_task() for managed-node3/include_tasks 13731 1727203831.87738: worker is 1 (out of 1 available) 13731 1727203831.87750: exiting _queue_task() for managed-node3/include_tasks 13731 1727203831.87762: done queuing things up, now waiting for results queue to drain 13731 1727203831.87764: waiting for pending results... 13731 1727203831.87934: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13731 1727203831.88020: in run() - task 028d2410-947f-82dc-c122-000000000277 13731 1727203831.88032: variable 'ansible_search_path' from source: unknown 13731 1727203831.88036: variable 'ansible_search_path' from source: unknown 13731 1727203831.88065: calling self._execute() 13731 1727203831.88131: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.88135: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.88143: variable 'omit' from source: magic vars 13731 1727203831.88409: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.88419: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.88423: _execute() done 13731 1727203831.88427: dumping result to json 13731 1727203831.88430: done dumping result, returning 13731 1727203831.88441: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role [028d2410-947f-82dc-c122-000000000277] 13731 1727203831.88443: sending task result for task 028d2410-947f-82dc-c122-000000000277 13731 1727203831.88524: done sending task result for task 028d2410-947f-82dc-c122-000000000277 13731 1727203831.88526: WORKER PROCESS EXITING 13731 1727203831.88577: no more pending results, returning what we have 13731 1727203831.88583: in VariableManager get_vars() 13731 1727203831.88622: Calling all_inventory to load vars for managed-node3 13731 1727203831.88625: Calling groups_inventory to load vars for managed-node3 13731 1727203831.88627: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.88638: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.88640: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.88643: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.88786: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.88924: done with get_vars() 13731 1727203831.88929: variable 'ansible_search_path' from source: unknown 13731 1727203831.88930: variable 'ansible_search_path' from source: unknown 13731 1727203831.88957: we have included files to process 13731 1727203831.88958: generating all_blocks data 13731 1727203831.88960: done generating all_blocks data 13731 1727203831.88964: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203831.88965: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203831.88967: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203831.89414: done processing included file 13731 1727203831.89415: iterating over new_blocks loaded from include file 13731 1727203831.89416: in VariableManager get_vars() 13731 1727203831.89443: done with get_vars() 13731 1727203831.89445: filtering new block on tags 13731 1727203831.89479: done filtering new block on tags 13731 1727203831.89482: in VariableManager get_vars() 13731 1727203831.89504: done with get_vars() 13731 1727203831.89506: filtering new block on tags 13731 1727203831.89546: done filtering new block on tags 13731 1727203831.89549: in VariableManager get_vars() 13731 1727203831.89571: done with get_vars() 13731 1727203831.89573: filtering new block on tags 13731 1727203831.89646: done filtering new block on tags 13731 1727203831.89648: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml for managed-node3 13731 1727203831.89653: extending task lists for all hosts with included blocks 13731 1727203831.91374: done extending task lists 13731 1727203831.91377: done processing included files 13731 1727203831.91378: results queue empty 13731 1727203831.91379: checking for any_errors_fatal 13731 1727203831.91381: done checking for any_errors_fatal 13731 1727203831.91382: checking for max_fail_percentage 13731 1727203831.91383: done checking for max_fail_percentage 13731 1727203831.91384: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.91384: done checking to see if all hosts have failed 13731 1727203831.91385: getting the remaining hosts for this loop 13731 1727203831.91387: done getting the remaining hosts for this loop 13731 1727203831.91389: getting the next task for host managed-node3 13731 1727203831.91393: done getting next task for host managed-node3 13731 1727203831.91396: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13731 1727203831.91400: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.91409: getting variables 13731 1727203831.91410: in VariableManager get_vars() 13731 1727203831.91422: Calling all_inventory to load vars for managed-node3 13731 1727203831.91424: Calling groups_inventory to load vars for managed-node3 13731 1727203831.91426: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.91431: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.91434: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.91436: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.91582: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.91782: done with get_vars() 13731 1727203831.91791: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role are present] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:3 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.043) 0:00:10.149 ***** 13731 1727203831.91860: entering _queue_task() for managed-node3/setup 13731 1727203831.92140: worker is 1 (out of 1 available) 13731 1727203831.92153: exiting _queue_task() for managed-node3/setup 13731 1727203831.92168: done queuing things up, now waiting for results queue to drain 13731 1727203831.92170: waiting for pending results... 13731 1727203831.92493: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13731 1727203831.92582: in run() - task 028d2410-947f-82dc-c122-0000000002d4 13731 1727203831.92590: variable 'ansible_search_path' from source: unknown 13731 1727203831.92611: variable 'ansible_search_path' from source: unknown 13731 1727203831.92637: calling self._execute() 13731 1727203831.92781: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.92784: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.92787: variable 'omit' from source: magic vars 13731 1727203831.93098: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.93114: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.93364: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203831.95535: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203831.95625: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203831.95880: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203831.95884: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203831.95886: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203831.95889: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203831.95891: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203831.95893: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203831.95933: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203831.95953: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203831.96022: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203831.96050: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203831.96085: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203831.96136: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203831.96156: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203831.96337: variable '__network_required_facts' from source: role '' defaults 13731 1727203831.96356: variable 'ansible_facts' from source: unknown 13731 1727203831.96452: Evaluated conditional (__network_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 13731 1727203831.96460: when evaluation is False, skipping this task 13731 1727203831.96469: _execute() done 13731 1727203831.96476: dumping result to json 13731 1727203831.96483: done dumping result, returning 13731 1727203831.96493: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present [028d2410-947f-82dc-c122-0000000002d4] 13731 1727203831.96500: sending task result for task 028d2410-947f-82dc-c122-0000000002d4 skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203831.96747: no more pending results, returning what we have 13731 1727203831.96751: results queue empty 13731 1727203831.96752: checking for any_errors_fatal 13731 1727203831.96754: done checking for any_errors_fatal 13731 1727203831.96755: checking for max_fail_percentage 13731 1727203831.96756: done checking for max_fail_percentage 13731 1727203831.96757: checking to see if all hosts have failed and the running result is not ok 13731 1727203831.96757: done checking to see if all hosts have failed 13731 1727203831.96758: getting the remaining hosts for this loop 13731 1727203831.96760: done getting the remaining hosts for this loop 13731 1727203831.96766: getting the next task for host managed-node3 13731 1727203831.96780: done getting next task for host managed-node3 13731 1727203831.96784: ^ task is: TASK: fedora.linux_system_roles.network : Check if system is ostree 13731 1727203831.96790: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203831.96803: getting variables 13731 1727203831.96805: in VariableManager get_vars() 13731 1727203831.96842: Calling all_inventory to load vars for managed-node3 13731 1727203831.96845: Calling groups_inventory to load vars for managed-node3 13731 1727203831.96847: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203831.96856: Calling all_plugins_play to load vars for managed-node3 13731 1727203831.96859: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203831.96864: Calling groups_plugins_play to load vars for managed-node3 13731 1727203831.97268: done sending task result for task 028d2410-947f-82dc-c122-0000000002d4 13731 1727203831.97278: WORKER PROCESS EXITING 13731 1727203831.97302: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203831.97520: done with get_vars() 13731 1727203831.97536: done getting variables TASK [fedora.linux_system_roles.network : Check if system is ostree] *********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:12 Tuesday 24 September 2024 14:50:31 -0400 (0:00:00.057) 0:00:10.207 ***** 13731 1727203831.97640: entering _queue_task() for managed-node3/stat 13731 1727203831.97908: worker is 1 (out of 1 available) 13731 1727203831.97920: exiting _queue_task() for managed-node3/stat 13731 1727203831.97931: done queuing things up, now waiting for results queue to drain 13731 1727203831.97933: waiting for pending results... 13731 1727203831.98212: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if system is ostree 13731 1727203831.98366: in run() - task 028d2410-947f-82dc-c122-0000000002d6 13731 1727203831.98399: variable 'ansible_search_path' from source: unknown 13731 1727203831.98516: variable 'ansible_search_path' from source: unknown 13731 1727203831.98520: calling self._execute() 13731 1727203831.98529: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203831.98542: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203831.98556: variable 'omit' from source: magic vars 13731 1727203831.98934: variable 'ansible_distribution_major_version' from source: facts 13731 1727203831.98957: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203831.99171: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203831.99408: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203831.99456: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203831.99505: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203831.99543: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203831.99670: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203831.99703: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203831.99742: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203831.99825: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203831.99932: variable '__network_is_ostree' from source: set_fact 13731 1727203831.99937: Evaluated conditional (not __network_is_ostree is defined): False 13731 1727203831.99940: when evaluation is False, skipping this task 13731 1727203831.99942: _execute() done 13731 1727203831.99944: dumping result to json 13731 1727203831.99947: done dumping result, returning 13731 1727203831.99949: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if system is ostree [028d2410-947f-82dc-c122-0000000002d6] 13731 1727203831.99951: sending task result for task 028d2410-947f-82dc-c122-0000000002d6 13731 1727203832.00157: done sending task result for task 028d2410-947f-82dc-c122-0000000002d6 13731 1727203832.00164: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13731 1727203832.00218: no more pending results, returning what we have 13731 1727203832.00222: results queue empty 13731 1727203832.00223: checking for any_errors_fatal 13731 1727203832.00232: done checking for any_errors_fatal 13731 1727203832.00233: checking for max_fail_percentage 13731 1727203832.00234: done checking for max_fail_percentage 13731 1727203832.00235: checking to see if all hosts have failed and the running result is not ok 13731 1727203832.00236: done checking to see if all hosts have failed 13731 1727203832.00237: getting the remaining hosts for this loop 13731 1727203832.00239: done getting the remaining hosts for this loop 13731 1727203832.00242: getting the next task for host managed-node3 13731 1727203832.00251: done getting next task for host managed-node3 13731 1727203832.00370: ^ task is: TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13731 1727203832.00379: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203832.00393: getting variables 13731 1727203832.00395: in VariableManager get_vars() 13731 1727203832.00426: Calling all_inventory to load vars for managed-node3 13731 1727203832.00429: Calling groups_inventory to load vars for managed-node3 13731 1727203832.00432: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203832.00440: Calling all_plugins_play to load vars for managed-node3 13731 1727203832.00442: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203832.00446: Calling groups_plugins_play to load vars for managed-node3 13731 1727203832.00639: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203832.01025: done with get_vars() 13731 1727203832.01036: done getting variables 13731 1727203832.01098: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Set flag to indicate system is ostree] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:17 Tuesday 24 September 2024 14:50:32 -0400 (0:00:00.034) 0:00:10.242 ***** 13731 1727203832.01132: entering _queue_task() for managed-node3/set_fact 13731 1727203832.01526: worker is 1 (out of 1 available) 13731 1727203832.01537: exiting _queue_task() for managed-node3/set_fact 13731 1727203832.01548: done queuing things up, now waiting for results queue to drain 13731 1727203832.01549: waiting for pending results... 13731 1727203832.01755: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13731 1727203832.01827: in run() - task 028d2410-947f-82dc-c122-0000000002d7 13731 1727203832.01851: variable 'ansible_search_path' from source: unknown 13731 1727203832.01861: variable 'ansible_search_path' from source: unknown 13731 1727203832.01903: calling self._execute() 13731 1727203832.01994: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203832.02005: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203832.02017: variable 'omit' from source: magic vars 13731 1727203832.02447: variable 'ansible_distribution_major_version' from source: facts 13731 1727203832.02462: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203832.02628: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203832.02895: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203832.02946: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203832.03050: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203832.03055: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203832.03112: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203832.03140: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203832.03178: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203832.03211: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203832.03303: variable '__network_is_ostree' from source: set_fact 13731 1727203832.03379: Evaluated conditional (not __network_is_ostree is defined): False 13731 1727203832.03382: when evaluation is False, skipping this task 13731 1727203832.03384: _execute() done 13731 1727203832.03386: dumping result to json 13731 1727203832.03389: done dumping result, returning 13731 1727203832.03392: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree [028d2410-947f-82dc-c122-0000000002d7] 13731 1727203832.03394: sending task result for task 028d2410-947f-82dc-c122-0000000002d7 13731 1727203832.03456: done sending task result for task 028d2410-947f-82dc-c122-0000000002d7 13731 1727203832.03459: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13731 1727203832.03507: no more pending results, returning what we have 13731 1727203832.03511: results queue empty 13731 1727203832.03512: checking for any_errors_fatal 13731 1727203832.03518: done checking for any_errors_fatal 13731 1727203832.03518: checking for max_fail_percentage 13731 1727203832.03520: done checking for max_fail_percentage 13731 1727203832.03521: checking to see if all hosts have failed and the running result is not ok 13731 1727203832.03522: done checking to see if all hosts have failed 13731 1727203832.03523: getting the remaining hosts for this loop 13731 1727203832.03525: done getting the remaining hosts for this loop 13731 1727203832.03530: getting the next task for host managed-node3 13731 1727203832.03540: done getting next task for host managed-node3 13731 1727203832.03544: ^ task is: TASK: fedora.linux_system_roles.network : Check which services are running 13731 1727203832.03550: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203832.03563: getting variables 13731 1727203832.03565: in VariableManager get_vars() 13731 1727203832.03714: Calling all_inventory to load vars for managed-node3 13731 1727203832.03717: Calling groups_inventory to load vars for managed-node3 13731 1727203832.03720: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203832.03729: Calling all_plugins_play to load vars for managed-node3 13731 1727203832.03732: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203832.03735: Calling groups_plugins_play to load vars for managed-node3 13731 1727203832.04054: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203832.04253: done with get_vars() 13731 1727203832.04264: done getting variables TASK [fedora.linux_system_roles.network : Check which services are running] **** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 Tuesday 24 September 2024 14:50:32 -0400 (0:00:00.032) 0:00:10.274 ***** 13731 1727203832.04364: entering _queue_task() for managed-node3/service_facts 13731 1727203832.04366: Creating lock for service_facts 13731 1727203832.04808: worker is 1 (out of 1 available) 13731 1727203832.04819: exiting _queue_task() for managed-node3/service_facts 13731 1727203832.04828: done queuing things up, now waiting for results queue to drain 13731 1727203832.04830: waiting for pending results... 13731 1727203832.05066: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which services are running 13731 1727203832.05071: in run() - task 028d2410-947f-82dc-c122-0000000002d9 13731 1727203832.05074: variable 'ansible_search_path' from source: unknown 13731 1727203832.05083: variable 'ansible_search_path' from source: unknown 13731 1727203832.05126: calling self._execute() 13731 1727203832.05222: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203832.05234: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203832.05248: variable 'omit' from source: magic vars 13731 1727203832.05613: variable 'ansible_distribution_major_version' from source: facts 13731 1727203832.05629: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203832.05641: variable 'omit' from source: magic vars 13731 1727203832.05716: variable 'omit' from source: magic vars 13731 1727203832.05782: variable 'omit' from source: magic vars 13731 1727203832.05805: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203832.05856: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203832.05925: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203832.05928: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203832.05930: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203832.05945: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203832.05952: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203832.05962: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203832.06062: Set connection var ansible_pipelining to False 13731 1727203832.06079: Set connection var ansible_shell_type to sh 13731 1727203832.06089: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203832.06098: Set connection var ansible_connection to ssh 13731 1727203832.06106: Set connection var ansible_shell_executable to /bin/sh 13731 1727203832.06142: Set connection var ansible_timeout to 10 13731 1727203832.06145: variable 'ansible_shell_executable' from source: unknown 13731 1727203832.06147: variable 'ansible_connection' from source: unknown 13731 1727203832.06152: variable 'ansible_module_compression' from source: unknown 13731 1727203832.06158: variable 'ansible_shell_type' from source: unknown 13731 1727203832.06164: variable 'ansible_shell_executable' from source: unknown 13731 1727203832.06170: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203832.06183: variable 'ansible_pipelining' from source: unknown 13731 1727203832.06252: variable 'ansible_timeout' from source: unknown 13731 1727203832.06256: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203832.06402: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203832.06417: variable 'omit' from source: magic vars 13731 1727203832.06425: starting attempt loop 13731 1727203832.06433: running the handler 13731 1727203832.06451: _low_level_execute_command(): starting 13731 1727203832.06462: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203832.07211: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203832.07235: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203832.07274: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203832.07335: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203832.07350: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203832.07399: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203832.07426: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203832.07452: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203832.07516: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203832.09188: stdout chunk (state=3): >>>/root <<< 13731 1727203832.09328: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203832.09348: stderr chunk (state=3): >>><<< 13731 1727203832.09358: stdout chunk (state=3): >>><<< 13731 1727203832.09401: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203832.09420: _low_level_execute_command(): starting 13731 1727203832.09432: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686 `" && echo ansible-tmp-1727203832.0940773-14493-117458920577686="` echo /root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686 `" ) && sleep 0' 13731 1727203832.10347: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203832.10400: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203832.10404: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203832.10444: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203832.10485: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203832.12379: stdout chunk (state=3): >>>ansible-tmp-1727203832.0940773-14493-117458920577686=/root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686 <<< 13731 1727203832.12500: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203832.12547: stderr chunk (state=3): >>><<< 13731 1727203832.12550: stdout chunk (state=3): >>><<< 13731 1727203832.12770: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203832.0940773-14493-117458920577686=/root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203832.12773: variable 'ansible_module_compression' from source: unknown 13731 1727203832.12777: ANSIBALLZ: Using lock for service_facts 13731 1727203832.12780: ANSIBALLZ: Acquiring lock 13731 1727203832.12782: ANSIBALLZ: Lock acquired: 140078454432416 13731 1727203832.12785: ANSIBALLZ: Creating module 13731 1727203832.26021: ANSIBALLZ: Writing module into payload 13731 1727203832.26084: ANSIBALLZ: Writing module 13731 1727203832.26106: ANSIBALLZ: Renaming module 13731 1727203832.26112: ANSIBALLZ: Done creating module 13731 1727203832.26127: variable 'ansible_facts' from source: unknown 13731 1727203832.26171: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/AnsiballZ_service_facts.py 13731 1727203832.26273: Sending initial data 13731 1727203832.26278: Sent initial data (162 bytes) 13731 1727203832.26713: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203832.26717: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203832.26738: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203832.26791: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203832.26810: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203832.26910: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203832.28512: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203832.28541: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203832.28577: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp5nmun5h0 /root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/AnsiballZ_service_facts.py <<< 13731 1727203832.28580: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/AnsiballZ_service_facts.py" <<< 13731 1727203832.28610: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp5nmun5h0" to remote "/root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/AnsiballZ_service_facts.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/AnsiballZ_service_facts.py" <<< 13731 1727203832.29133: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203832.29249: stderr chunk (state=3): >>><<< 13731 1727203832.29252: stdout chunk (state=3): >>><<< 13731 1727203832.29254: done transferring module to remote 13731 1727203832.29256: _low_level_execute_command(): starting 13731 1727203832.29258: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/ /root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/AnsiballZ_service_facts.py && sleep 0' 13731 1727203832.29609: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203832.29629: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203832.29667: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203832.29702: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203832.29735: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203832.29788: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203832.31554: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203832.31565: stdout chunk (state=3): >>><<< 13731 1727203832.31579: stderr chunk (state=3): >>><<< 13731 1727203832.31679: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203832.31683: _low_level_execute_command(): starting 13731 1727203832.31685: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/AnsiballZ_service_facts.py && sleep 0' 13731 1727203832.32202: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203832.32216: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203832.32229: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203832.32244: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203832.32341: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203832.32366: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203832.32434: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203833.83593: stdout chunk (state=3): >>> {"ansible_facts": {"services": {"audit-rules.service": {"name": "audit-rules.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dm-event.service": {"name": "dm-event.service", "state": "stopped", "status": "static", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "st<<< 13731 1727203833.83618: stdout chunk (state=3): >>>opped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-lvmpolld.service": {"name": "lvm2-lvmpolld.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-monitor.service": {"name": "lvm2-monitor.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@dm_mod.service": {"name": "modprobe@dm_mod.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@efi_pstore.service": {"name": "modprobe@efi_pstore.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@loop.service": {"name": "modprobe@loop.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "pcscd.service": {"name": "pcscd.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-ma<<< 13731 1727203833.83630: stdout chunk (state=3): >>>rk.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ssh-host-keys-migration.service": {"name": "ssh-host-keys-migration.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-battery-check.service": {"name": "systemd-battery-check.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-confext.service": {"name": "systemd-confext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-clear.service": {"name": "systemd-hibernate-clear.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-resume.service": {"name": "systemd-hibernate-resume.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.s<<< 13731 1727203833.83638: stdout chunk (state=3): >>>ervice", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-oomd.service": {"name": "systemd-oomd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-quotacheck-root.service": {"name": "systemd-quotacheck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-soft-reboot.service": {"name": "systemd-soft-reboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev-early.service": {"name": "systemd-tmpfiles-setup-dev-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup-early.service": {"name": "systemd-tpm2-setup-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup.service": {"name": "systemd-tpm2-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-load-credentials.service": {"name": "systemd-udev-load-credentials.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "stat<<< 13731 1727203833.83665: stdout chunk (state=3): >>>us": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "blk-availability.service": {"name": "blk-availability.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "capsule@.service": {"name": "capsule@.service", "state": "unknown", "status": "static", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd.service": {"name": "dhcpcd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd@.service": {"name": "dhcpcd@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fips-crypto-policy-overlay.service": {"name": "fips-crypto-policy-overlay.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fsidd.service": {"name": "fsidd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive",<<< 13731 1727203833.83686: stdout chunk (state=3): >>> "status": "disabled", "source": "systemd"}, "lvm-devices-import.service": {"name": "lvm-devices-import.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "polkit.service": {"name": "polkit.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon-root.service": {"name": "quotaon-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "quotaon@.service": {"name": "quotaon@.service", "state": "unknown", "status": "static", "source": "systemd"}, "rpmdb-migrate.service": {"name": "rpmdb-migrate.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-boot-update.service": {"name": "syste<<< 13731 1727203833.83693: stdout chunk (state=3): >>>md-boot-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-bootctl@.service": {"name": "systemd-bootctl@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-creds@.service": {"name": "systemd-creds@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald-sync@.service": {"name": "systemd-journald-sync@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrextend@.service": {"name": "systemd-pcrextend@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrlock-file-system.service": {"name": "systemd-pcrlock-file-system.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-code.service": {"name": "systemd-pcrlock-firmware-code.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-config.service": {"name": "systemd-pcrlock-firmware-config.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-machine-id.service": {"name": "systemd-pcrlock-machine-id.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-make-policy.service": {"name": "systemd-pcrlock-make-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-authority.service": {"name": "systemd-pcrlock-secureboot-authority.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-policy.service": {"name": "systemd-pcrlock-secureboot-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock@.service": {"name": "systemd-pcrlock@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-quotacheck@.service": {"name": "systemd-quotacheck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysext@.service": {"name": "systemd-sysext@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} <<< 13731 1727203833.85174: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203833.85219: stderr chunk (state=3): >>><<< 13731 1727203833.85222: stdout chunk (state=3): >>><<< 13731 1727203833.85385: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"services": {"audit-rules.service": {"name": "audit-rules.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dm-event.service": {"name": "dm-event.service", "state": "stopped", "status": "static", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-lvmpolld.service": {"name": "lvm2-lvmpolld.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-monitor.service": {"name": "lvm2-monitor.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@dm_mod.service": {"name": "modprobe@dm_mod.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@efi_pstore.service": {"name": "modprobe@efi_pstore.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@loop.service": {"name": "modprobe@loop.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "pcscd.service": {"name": "pcscd.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ssh-host-keys-migration.service": {"name": "ssh-host-keys-migration.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-battery-check.service": {"name": "systemd-battery-check.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-confext.service": {"name": "systemd-confext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-clear.service": {"name": "systemd-hibernate-clear.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-resume.service": {"name": "systemd-hibernate-resume.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-oomd.service": {"name": "systemd-oomd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-quotacheck-root.service": {"name": "systemd-quotacheck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-soft-reboot.service": {"name": "systemd-soft-reboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev-early.service": {"name": "systemd-tmpfiles-setup-dev-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup-early.service": {"name": "systemd-tpm2-setup-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup.service": {"name": "systemd-tpm2-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-load-credentials.service": {"name": "systemd-udev-load-credentials.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "blk-availability.service": {"name": "blk-availability.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "capsule@.service": {"name": "capsule@.service", "state": "unknown", "status": "static", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd.service": {"name": "dhcpcd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd@.service": {"name": "dhcpcd@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fips-crypto-policy-overlay.service": {"name": "fips-crypto-policy-overlay.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fsidd.service": {"name": "fsidd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "lvm-devices-import.service": {"name": "lvm-devices-import.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "polkit.service": {"name": "polkit.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon-root.service": {"name": "quotaon-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "quotaon@.service": {"name": "quotaon@.service", "state": "unknown", "status": "static", "source": "systemd"}, "rpmdb-migrate.service": {"name": "rpmdb-migrate.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-bootctl@.service": {"name": "systemd-bootctl@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-creds@.service": {"name": "systemd-creds@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald-sync@.service": {"name": "systemd-journald-sync@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrextend@.service": {"name": "systemd-pcrextend@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrlock-file-system.service": {"name": "systemd-pcrlock-file-system.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-code.service": {"name": "systemd-pcrlock-firmware-code.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-config.service": {"name": "systemd-pcrlock-firmware-config.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-machine-id.service": {"name": "systemd-pcrlock-machine-id.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-make-policy.service": {"name": "systemd-pcrlock-make-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-authority.service": {"name": "systemd-pcrlock-secureboot-authority.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-policy.service": {"name": "systemd-pcrlock-secureboot-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock@.service": {"name": "systemd-pcrlock@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-quotacheck@.service": {"name": "systemd-quotacheck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysext@.service": {"name": "systemd-sysext@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203833.85933: done with _execute_module (service_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'service_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203833.85960: _low_level_execute_command(): starting 13731 1727203833.85970: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203832.0940773-14493-117458920577686/ > /dev/null 2>&1 && sleep 0' 13731 1727203833.86455: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203833.86461: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203833.86494: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203833.86497: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203833.86499: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203833.86501: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203833.86556: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203833.86563: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203833.86565: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203833.86599: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203833.88454: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203833.88458: stdout chunk (state=3): >>><<< 13731 1727203833.88460: stderr chunk (state=3): >>><<< 13731 1727203833.88680: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203833.88684: handler run complete 13731 1727203833.88686: variable 'ansible_facts' from source: unknown 13731 1727203833.88878: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203833.89165: variable 'ansible_facts' from source: unknown 13731 1727203833.90021: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203833.90135: attempt loop complete, returning result 13731 1727203833.90139: _execute() done 13731 1727203833.90141: dumping result to json 13731 1727203833.90179: done dumping result, returning 13731 1727203833.90187: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which services are running [028d2410-947f-82dc-c122-0000000002d9] 13731 1727203833.90191: sending task result for task 028d2410-947f-82dc-c122-0000000002d9 ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203833.90731: no more pending results, returning what we have 13731 1727203833.90733: results queue empty 13731 1727203833.90734: checking for any_errors_fatal 13731 1727203833.90738: done checking for any_errors_fatal 13731 1727203833.90739: checking for max_fail_percentage 13731 1727203833.90740: done checking for max_fail_percentage 13731 1727203833.90741: checking to see if all hosts have failed and the running result is not ok 13731 1727203833.90742: done checking to see if all hosts have failed 13731 1727203833.90742: getting the remaining hosts for this loop 13731 1727203833.90744: done getting the remaining hosts for this loop 13731 1727203833.90747: getting the next task for host managed-node3 13731 1727203833.90753: done getting next task for host managed-node3 13731 1727203833.90756: ^ task is: TASK: fedora.linux_system_roles.network : Check which packages are installed 13731 1727203833.90761: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203833.90770: getting variables 13731 1727203833.90771: in VariableManager get_vars() 13731 1727203833.90808: Calling all_inventory to load vars for managed-node3 13731 1727203833.90811: Calling groups_inventory to load vars for managed-node3 13731 1727203833.90813: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203833.90818: done sending task result for task 028d2410-947f-82dc-c122-0000000002d9 13731 1727203833.90821: WORKER PROCESS EXITING 13731 1727203833.90828: Calling all_plugins_play to load vars for managed-node3 13731 1727203833.90833: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203833.90835: Calling groups_plugins_play to load vars for managed-node3 13731 1727203833.91066: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203833.91344: done with get_vars() 13731 1727203833.91353: done getting variables TASK [fedora.linux_system_roles.network : Check which packages are installed] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 Tuesday 24 September 2024 14:50:33 -0400 (0:00:01.870) 0:00:12.145 ***** 13731 1727203833.91420: entering _queue_task() for managed-node3/package_facts 13731 1727203833.91421: Creating lock for package_facts 13731 1727203833.91629: worker is 1 (out of 1 available) 13731 1727203833.91642: exiting _queue_task() for managed-node3/package_facts 13731 1727203833.91654: done queuing things up, now waiting for results queue to drain 13731 1727203833.91656: waiting for pending results... 13731 1727203833.91819: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which packages are installed 13731 1727203833.91913: in run() - task 028d2410-947f-82dc-c122-0000000002da 13731 1727203833.91925: variable 'ansible_search_path' from source: unknown 13731 1727203833.91929: variable 'ansible_search_path' from source: unknown 13731 1727203833.91955: calling self._execute() 13731 1727203833.92019: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203833.92023: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203833.92033: variable 'omit' from source: magic vars 13731 1727203833.92290: variable 'ansible_distribution_major_version' from source: facts 13731 1727203833.92299: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203833.92305: variable 'omit' from source: magic vars 13731 1727203833.92354: variable 'omit' from source: magic vars 13731 1727203833.92380: variable 'omit' from source: magic vars 13731 1727203833.92410: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203833.92437: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203833.92451: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203833.92469: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203833.92483: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203833.92505: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203833.92509: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203833.92511: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203833.92582: Set connection var ansible_pipelining to False 13731 1727203833.92586: Set connection var ansible_shell_type to sh 13731 1727203833.92591: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203833.92596: Set connection var ansible_connection to ssh 13731 1727203833.92601: Set connection var ansible_shell_executable to /bin/sh 13731 1727203833.92606: Set connection var ansible_timeout to 10 13731 1727203833.92621: variable 'ansible_shell_executable' from source: unknown 13731 1727203833.92624: variable 'ansible_connection' from source: unknown 13731 1727203833.92627: variable 'ansible_module_compression' from source: unknown 13731 1727203833.92629: variable 'ansible_shell_type' from source: unknown 13731 1727203833.92632: variable 'ansible_shell_executable' from source: unknown 13731 1727203833.92634: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203833.92638: variable 'ansible_pipelining' from source: unknown 13731 1727203833.92640: variable 'ansible_timeout' from source: unknown 13731 1727203833.92643: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203833.92783: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203833.92791: variable 'omit' from source: magic vars 13731 1727203833.92796: starting attempt loop 13731 1727203833.92799: running the handler 13731 1727203833.92810: _low_level_execute_command(): starting 13731 1727203833.92817: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203833.93325: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203833.93329: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203833.93332: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203833.93385: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203833.93389: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203833.93437: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203833.95037: stdout chunk (state=3): >>>/root <<< 13731 1727203833.95132: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203833.95161: stderr chunk (state=3): >>><<< 13731 1727203833.95164: stdout chunk (state=3): >>><<< 13731 1727203833.95185: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203833.95196: _low_level_execute_command(): starting 13731 1727203833.95201: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290 `" && echo ansible-tmp-1727203833.951846-14552-196722913906290="` echo /root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290 `" ) && sleep 0' 13731 1727203833.95628: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203833.95631: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203833.95633: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203833.95642: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203833.95645: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203833.95682: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203833.95694: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203833.95726: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203833.97602: stdout chunk (state=3): >>>ansible-tmp-1727203833.951846-14552-196722913906290=/root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290 <<< 13731 1727203833.97710: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203833.97735: stderr chunk (state=3): >>><<< 13731 1727203833.97738: stdout chunk (state=3): >>><<< 13731 1727203833.97751: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203833.951846-14552-196722913906290=/root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203833.97791: variable 'ansible_module_compression' from source: unknown 13731 1727203833.97828: ANSIBALLZ: Using lock for package_facts 13731 1727203833.97831: ANSIBALLZ: Acquiring lock 13731 1727203833.97834: ANSIBALLZ: Lock acquired: 140078449960976 13731 1727203833.97836: ANSIBALLZ: Creating module 13731 1727203834.28493: ANSIBALLZ: Writing module into payload 13731 1727203834.28647: ANSIBALLZ: Writing module 13731 1727203834.28681: ANSIBALLZ: Renaming module 13731 1727203834.28716: ANSIBALLZ: Done creating module 13731 1727203834.28924: variable 'ansible_facts' from source: unknown 13731 1727203834.29251: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/AnsiballZ_package_facts.py 13731 1727203834.29488: Sending initial data 13731 1727203834.29498: Sent initial data (161 bytes) 13731 1727203834.30507: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203834.30522: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203834.30533: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203834.30791: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203834.30825: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203834.31006: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203834.32697: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203834.32725: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpre7hr7zn /root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/AnsiballZ_package_facts.py <<< 13731 1727203834.32728: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/AnsiballZ_package_facts.py" <<< 13731 1727203834.32748: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpre7hr7zn" to remote "/root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/AnsiballZ_package_facts.py" <<< 13731 1727203834.32762: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/AnsiballZ_package_facts.py" <<< 13731 1727203834.35504: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203834.35784: stderr chunk (state=3): >>><<< 13731 1727203834.35788: stdout chunk (state=3): >>><<< 13731 1727203834.35790: done transferring module to remote 13731 1727203834.35792: _low_level_execute_command(): starting 13731 1727203834.35794: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/ /root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/AnsiballZ_package_facts.py && sleep 0' 13731 1727203834.36941: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203834.37091: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203834.37150: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203834.37243: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203834.37258: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203834.37320: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203834.39168: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203834.39184: stdout chunk (state=3): >>><<< 13731 1727203834.39196: stderr chunk (state=3): >>><<< 13731 1727203834.39372: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203834.39378: _low_level_execute_command(): starting 13731 1727203834.39382: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/AnsiballZ_package_facts.py && sleep 0' 13731 1727203834.40434: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203834.40593: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203834.40816: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203834.40844: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203834.84500: stdout chunk (state=3): >>> {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "hunspell-filesystem": [{"name": "hunspell-filesystem", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "google-noto-fonts-common": [{"name": "google-noto-fonts-common", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-mono-vf-fonts": [{"name": "google-noto-sans-mono-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-vf-fonts": [{"name": "google-noto-sans-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-serif-vf-fonts": [{"name": "google-noto-serif-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-mono-vf-fonts": [{"name": "redhat-mono-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-text-vf-fonts": [{"name": "redhat-text-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "default-fonts-core-sans": [{"name": "default-fonts-core-sans", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-fonts-en": [{"name": "langpacks-fonts-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-ucode-firmware": [{"name": "amd-ucode-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "atheros-firmware": [{"name": "atheros-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "brcmfmac-firmware": [{"name": "brcmfmac-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "cirrus-audio-firmware": [{"name": "cirrus-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-audio-firmware": [{"name": "intel-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "mt7xxx-firmware": [{"name": "mt7xxx-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nxpwireless-firmware": [{"name": "nxpwireless-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "realtek-firmware": [{"name": "realtek-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tiwilink-firmware": [{"name": "tiwilink-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-gpu-firmware": [{"name": "amd-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-gpu-firmware": [{"name": "intel-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nvidia-gpu-firmware": [{"name": "nvidia-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "xkeyboard-config": [{"name": "xkeyboard-config", "version": "2.41", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks"<<< 13731 1727203834.84542: stdout chunk (state=3): >>>: [{"name": "gawk-all-langpacks", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-data": [{"name": "vim-data", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20240107", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.379", "release": "10.1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.14.5", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.18", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "21.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.2.26", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-ng-compat": [{"name": "zlib-ng-compat", "version": "2.1.6", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.19", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.5", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.5", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.36", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.30", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.50", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.11", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.46.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.4", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "1.1", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "51.20230828cvs.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.6.2", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.1", "release": "9.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.17", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.9", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.10.0", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.5.0", "release": "1.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "hunspell-en-GB": [{"name": "hunspell-en-GB", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell-en-US": [{"name": "hunspell-en-US", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell": [{"name": "hunspell", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.6.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam-libs": [{"name": "pam-libs", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.69", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.15.0", "release": "3.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.19.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.80.4", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-libs": [{"name": "polkit-libs", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "101.2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.13", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "31", "release": "8.el10", "epoch": null,<<< 13731 1727203834.84595: stdout chunk (state=3): >>> "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.11.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.6.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.7", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.10", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.3", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.6", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.4", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.7.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.16", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.6.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.5", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "42.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.7", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sequoia": [{"name": "rpm-sequoia", "version": "1.6.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.29", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.79.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsecret": [{"name": "libsecret", "version": "0.21.2", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pinentry": [{"name": "pinentry", "version": "1.3.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libusb1": [{"name": "libusb1", "version": "1.0.27", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "4.0.4", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hunspell-en": [{"name": "hunspell-en", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.2.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm": [{"name": "gdbm", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libxkbcommon": [{"name": "libxkbcommon", "version": "1.7.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.23.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.7", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.10", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.8", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "35", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "661", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.6", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "memstrack": [{"name": "memstrack", "version": "0.2.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.25.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.15", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "duktape": [{"name": "duktape", "version": "2.7.0", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "22.el10.gating_test1", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse3-libs": [{"name": "fuse3-libs", "version": "3.16.2", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-envsubst": [{"name": "gettext-envsubst", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-runtime": [{"name": "gettext-runtime", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "58", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.1.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.11.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.14.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "28.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.62.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.7", "release": "6.el10", "epoch": null, "arch": "x86_64", "source":<<< 13731 1727203834.84695: stdout chunk (state=3): >>> "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-gssapi": [{"name": "cyrus-sasl-gssapi", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11^20190603git9e74f2d", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "polkit": [{"name": "polkit", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-pkla-compat": [{"name": "polkit-pkla-compat", "version": "0.1", "release": "29.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "audit-rules": [{"name": "audit-rules", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20240905", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hasher": [{"name": "libkcapi-hasher", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.22.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "makedumpfile": [{"name": "makedumpfile", "version": "1.7.5", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.9.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.18.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss-fapi": [{"name": "tpm2-tss-fapi", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tools": [{"name": "tpm2-tools", "version": "5.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "76.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.81", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.15.0", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.32", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-krb5-common": [{"name": "sssd-krb5-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpdecimal": [{"name": "mpdecimal", "version": "2.5.1", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.3.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo-python-plugin": [{"name": "sudo-python-plugin", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.16.0", "release": "15.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.2", "release": "14.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "235", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.9", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.7.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.29", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kdump-utils": [{"name": "kdump-utils", "version": "1.0.43", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.13.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-libs": [{"name": "pcsc-lite-libs", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-ccid": [{"name": "pcsc-lite-ccid", "version": "1.6.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite": [{"name": "pcsc-lite", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2-smime": [{"name": "gnupg2-smime", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.3", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.24", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.14.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "1.2304.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2408.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "8.el10", "epoch": 18, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.12.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.2.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.20", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwlwifi-dvm-firmware": [{"name": "iwlwifi-dvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwlwifi-mvm-firmware": [{"name": "iwlwifi-mvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "37.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.5", "release": "0.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.20", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.59", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.89", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.189", "release": "511.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.15", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.27", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2024.001", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20231213", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.61.000", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-locale": [{"name": "perl-locale", "version": "1.12", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.42", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.350", "release": "510.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.085", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.68", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "2.20", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.22", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.45", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.088", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.038", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.09", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "5.01", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.18", "release": "510.el10", "epoch<<< 13731 1727203834.84741: stdout chunk (state=3): >>>": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.31", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.29", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.55", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.03", "release": "510.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.38", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.86", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.63", "release": "510.el10", "epoch": 5, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.32", "release": "510.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.37", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.241", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.58", "release": "2.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.54", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.78", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.91", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.56", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.17", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.21", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.44", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "23.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xxd": [{"name": "xxd", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.39", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "5.2.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "24.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.35", "release": "4.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.10", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.24.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.4.1", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.9.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.3.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpds-py": [{"name": "python3-rpds-py", "version": "0.17.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "23.2.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-referencing": [{"name": "python3-referencing", "version": "0.31.1", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.19", "release": "1.el10", "epoch<<< 13731 1727203834.84746: stdout chunk (state=3): >>>": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema-specifications": [{"name": "python3-jsonschema-specifications", "version": "2023.11.2", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "4.19.1", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.5", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.2.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "2.1.3", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "3.1.4", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.3", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.33", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.9.0", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.8", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "69.0.3", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.5.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "6.0.1", "release": "18.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-charset-normalizer": [{"name": "python3-charset-normalizer", "version": "3.3.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.32.3", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dhcpcd": [{"name": "dhcpcd", "version": "10.0.6", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "24.1.4", "release": "17.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "device-mapper-event-libs": [{"name": "device-mapper-event-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "libaio": [{"name": "libaio", "version": "0.3.111", "release": "20.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-event": [{"name": "device-mapper-event", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "lvm2-libs": [{"name": "lvm2-libs", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper-persistent-data": [{"name": "device-mapper-persistent-data", "version": "1.0.11", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lvm2": [{"name": "lvm2", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.17", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.90", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} <<< 13731 1727203834.86592: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203834.86596: stdout chunk (state=3): >>><<< 13731 1727203834.86598: stderr chunk (state=3): >>><<< 13731 1727203834.86741: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "hunspell-filesystem": [{"name": "hunspell-filesystem", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "google-noto-fonts-common": [{"name": "google-noto-fonts-common", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-mono-vf-fonts": [{"name": "google-noto-sans-mono-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-vf-fonts": [{"name": "google-noto-sans-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-serif-vf-fonts": [{"name": "google-noto-serif-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-mono-vf-fonts": [{"name": "redhat-mono-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-text-vf-fonts": [{"name": "redhat-text-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "default-fonts-core-sans": [{"name": "default-fonts-core-sans", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-fonts-en": [{"name": "langpacks-fonts-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-ucode-firmware": [{"name": "amd-ucode-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "atheros-firmware": [{"name": "atheros-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "brcmfmac-firmware": [{"name": "brcmfmac-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "cirrus-audio-firmware": [{"name": "cirrus-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-audio-firmware": [{"name": "intel-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "mt7xxx-firmware": [{"name": "mt7xxx-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nxpwireless-firmware": [{"name": "nxpwireless-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "realtek-firmware": [{"name": "realtek-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tiwilink-firmware": [{"name": "tiwilink-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-gpu-firmware": [{"name": "amd-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-gpu-firmware": [{"name": "intel-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nvidia-gpu-firmware": [{"name": "nvidia-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "xkeyboard-config": [{"name": "xkeyboard-config", "version": "2.41", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-data": [{"name": "vim-data", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20240107", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.379", "release": "10.1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.14.5", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.18", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "21.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.2.26", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-ng-compat": [{"name": "zlib-ng-compat", "version": "2.1.6", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.19", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.5", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.5", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.36", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.30", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.50", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.11", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.46.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.4", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "1.1", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "51.20230828cvs.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.6.2", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.1", "release": "9.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.17", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.9", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.10.0", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.5.0", "release": "1.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "hunspell-en-GB": [{"name": "hunspell-en-GB", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell-en-US": [{"name": "hunspell-en-US", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell": [{"name": "hunspell", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.6.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam-libs": [{"name": "pam-libs", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.69", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.15.0", "release": "3.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.19.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.80.4", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-libs": [{"name": "polkit-libs", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "101.2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.13", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.11.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.6.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.7", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.10", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.3", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.6", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.4", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.7.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.16", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.6.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.5", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "42.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.7", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sequoia": [{"name": "rpm-sequoia", "version": "1.6.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.29", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.79.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsecret": [{"name": "libsecret", "version": "0.21.2", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pinentry": [{"name": "pinentry", "version": "1.3.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libusb1": [{"name": "libusb1", "version": "1.0.27", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "4.0.4", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hunspell-en": [{"name": "hunspell-en", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.2.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm": [{"name": "gdbm", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libxkbcommon": [{"name": "libxkbcommon", "version": "1.7.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.23.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.7", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.10", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.8", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "35", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "661", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.6", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "memstrack": [{"name": "memstrack", "version": "0.2.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.25.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.15", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "duktape": [{"name": "duktape", "version": "2.7.0", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "22.el10.gating_test1", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse3-libs": [{"name": "fuse3-libs", "version": "3.16.2", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-envsubst": [{"name": "gettext-envsubst", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-runtime": [{"name": "gettext-runtime", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "58", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.1.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.11.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.14.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "28.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.62.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.7", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-gssapi": [{"name": "cyrus-sasl-gssapi", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11^20190603git9e74f2d", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "polkit": [{"name": "polkit", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-pkla-compat": [{"name": "polkit-pkla-compat", "version": "0.1", "release": "29.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "audit-rules": [{"name": "audit-rules", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20240905", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hasher": [{"name": "libkcapi-hasher", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.22.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "makedumpfile": [{"name": "makedumpfile", "version": "1.7.5", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.9.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.18.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss-fapi": [{"name": "tpm2-tss-fapi", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tools": [{"name": "tpm2-tools", "version": "5.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "76.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.81", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.15.0", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.32", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-krb5-common": [{"name": "sssd-krb5-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpdecimal": [{"name": "mpdecimal", "version": "2.5.1", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.3.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo-python-plugin": [{"name": "sudo-python-plugin", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.16.0", "release": "15.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.2", "release": "14.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "235", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.9", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.7.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.29", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kdump-utils": [{"name": "kdump-utils", "version": "1.0.43", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.13.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-libs": [{"name": "pcsc-lite-libs", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-ccid": [{"name": "pcsc-lite-ccid", "version": "1.6.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite": [{"name": "pcsc-lite", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2-smime": [{"name": "gnupg2-smime", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.3", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.24", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.14.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "1.2304.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2408.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "8.el10", "epoch": 18, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.12.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.2.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.20", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwlwifi-dvm-firmware": [{"name": "iwlwifi-dvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwlwifi-mvm-firmware": [{"name": "iwlwifi-mvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "37.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.5", "release": "0.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.20", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.59", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.89", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.189", "release": "511.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.15", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.27", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2024.001", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20231213", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.61.000", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-locale": [{"name": "perl-locale", "version": "1.12", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.42", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.350", "release": "510.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.085", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.68", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "2.20", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.22", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.45", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.088", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.038", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.09", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "5.01", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.18", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.31", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.29", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.55", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.03", "release": "510.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.38", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.86", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.63", "release": "510.el10", "epoch": 5, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.32", "release": "510.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.37", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.241", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.58", "release": "2.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.54", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.78", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.91", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.56", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.17", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.21", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.44", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "23.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xxd": [{"name": "xxd", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.39", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "5.2.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "24.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.35", "release": "4.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.10", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.24.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.4.1", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.9.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.3.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpds-py": [{"name": "python3-rpds-py", "version": "0.17.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "23.2.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-referencing": [{"name": "python3-referencing", "version": "0.31.1", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.19", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema-specifications": [{"name": "python3-jsonschema-specifications", "version": "2023.11.2", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "4.19.1", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.5", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.2.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "2.1.3", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "3.1.4", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.3", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.33", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.9.0", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.8", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "69.0.3", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.5.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "6.0.1", "release": "18.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-charset-normalizer": [{"name": "python3-charset-normalizer", "version": "3.3.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.32.3", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dhcpcd": [{"name": "dhcpcd", "version": "10.0.6", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "24.1.4", "release": "17.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "device-mapper-event-libs": [{"name": "device-mapper-event-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "libaio": [{"name": "libaio", "version": "0.3.111", "release": "20.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-event": [{"name": "device-mapper-event", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "lvm2-libs": [{"name": "lvm2-libs", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper-persistent-data": [{"name": "device-mapper-persistent-data", "version": "1.0.11", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lvm2": [{"name": "lvm2", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.17", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.90", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203834.88747: done with _execute_module (package_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'package_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203834.88788: _low_level_execute_command(): starting 13731 1727203834.88799: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203833.951846-14552-196722913906290/ > /dev/null 2>&1 && sleep 0' 13731 1727203834.89484: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203834.89500: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203834.89521: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203834.89636: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203834.89659: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203834.89678: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203834.89741: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203834.91616: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203834.91628: stdout chunk (state=3): >>><<< 13731 1727203834.91640: stderr chunk (state=3): >>><<< 13731 1727203834.91661: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203834.91679: handler run complete 13731 1727203834.92843: variable 'ansible_facts' from source: unknown 13731 1727203834.93261: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203834.95095: variable 'ansible_facts' from source: unknown 13731 1727203834.95681: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203834.96204: attempt loop complete, returning result 13731 1727203834.96224: _execute() done 13731 1727203834.96231: dumping result to json 13731 1727203834.96448: done dumping result, returning 13731 1727203834.96466: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which packages are installed [028d2410-947f-82dc-c122-0000000002da] 13731 1727203834.96480: sending task result for task 028d2410-947f-82dc-c122-0000000002da 13731 1727203835.03178: done sending task result for task 028d2410-947f-82dc-c122-0000000002da 13731 1727203835.03182: WORKER PROCESS EXITING ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203835.03284: no more pending results, returning what we have 13731 1727203835.03287: results queue empty 13731 1727203835.03288: checking for any_errors_fatal 13731 1727203835.03292: done checking for any_errors_fatal 13731 1727203835.03293: checking for max_fail_percentage 13731 1727203835.03294: done checking for max_fail_percentage 13731 1727203835.03295: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.03296: done checking to see if all hosts have failed 13731 1727203835.03297: getting the remaining hosts for this loop 13731 1727203835.03298: done getting the remaining hosts for this loop 13731 1727203835.03302: getting the next task for host managed-node3 13731 1727203835.03308: done getting next task for host managed-node3 13731 1727203835.03311: ^ task is: TASK: fedora.linux_system_roles.network : Print network provider 13731 1727203835.03316: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.03324: getting variables 13731 1727203835.03325: in VariableManager get_vars() 13731 1727203835.03347: Calling all_inventory to load vars for managed-node3 13731 1727203835.03349: Calling groups_inventory to load vars for managed-node3 13731 1727203835.03351: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.03358: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.03360: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.03365: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.04984: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.06841: done with get_vars() 13731 1727203835.06876: done getting variables 13731 1727203835.06937: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Print network provider] ************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:7 Tuesday 24 September 2024 14:50:35 -0400 (0:00:01.155) 0:00:13.300 ***** 13731 1727203835.06983: entering _queue_task() for managed-node3/debug 13731 1727203835.07312: worker is 1 (out of 1 available) 13731 1727203835.07325: exiting _queue_task() for managed-node3/debug 13731 1727203835.07338: done queuing things up, now waiting for results queue to drain 13731 1727203835.07339: waiting for pending results... 13731 1727203835.07708: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Print network provider 13731 1727203835.08084: in run() - task 028d2410-947f-82dc-c122-000000000278 13731 1727203835.08087: variable 'ansible_search_path' from source: unknown 13731 1727203835.08090: variable 'ansible_search_path' from source: unknown 13731 1727203835.08093: calling self._execute() 13731 1727203835.08146: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.08201: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.08216: variable 'omit' from source: magic vars 13731 1727203835.09003: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.09070: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.09083: variable 'omit' from source: magic vars 13731 1727203835.09279: variable 'omit' from source: magic vars 13731 1727203835.09366: variable 'network_provider' from source: set_fact 13731 1727203835.09502: variable 'omit' from source: magic vars 13731 1727203835.09547: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203835.09634: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203835.09726: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203835.09747: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203835.09765: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203835.09803: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203835.09887: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.09895: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.10004: Set connection var ansible_pipelining to False 13731 1727203835.10180: Set connection var ansible_shell_type to sh 13731 1727203835.10183: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203835.10185: Set connection var ansible_connection to ssh 13731 1727203835.10187: Set connection var ansible_shell_executable to /bin/sh 13731 1727203835.10190: Set connection var ansible_timeout to 10 13731 1727203835.10207: variable 'ansible_shell_executable' from source: unknown 13731 1727203835.10254: variable 'ansible_connection' from source: unknown 13731 1727203835.10265: variable 'ansible_module_compression' from source: unknown 13731 1727203835.10272: variable 'ansible_shell_type' from source: unknown 13731 1727203835.10280: variable 'ansible_shell_executable' from source: unknown 13731 1727203835.10287: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.10294: variable 'ansible_pipelining' from source: unknown 13731 1727203835.10355: variable 'ansible_timeout' from source: unknown 13731 1727203835.10358: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.10983: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203835.10987: variable 'omit' from source: magic vars 13731 1727203835.10989: starting attempt loop 13731 1727203835.10991: running the handler 13731 1727203835.10993: handler run complete 13731 1727203835.10995: attempt loop complete, returning result 13731 1727203835.10997: _execute() done 13731 1727203835.10999: dumping result to json 13731 1727203835.11001: done dumping result, returning 13731 1727203835.11003: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Print network provider [028d2410-947f-82dc-c122-000000000278] 13731 1727203835.11005: sending task result for task 028d2410-947f-82dc-c122-000000000278 13731 1727203835.11076: done sending task result for task 028d2410-947f-82dc-c122-000000000278 13731 1727203835.11080: WORKER PROCESS EXITING ok: [managed-node3] => {} MSG: Using network provider: nm 13731 1727203835.11141: no more pending results, returning what we have 13731 1727203835.11145: results queue empty 13731 1727203835.11145: checking for any_errors_fatal 13731 1727203835.11155: done checking for any_errors_fatal 13731 1727203835.11155: checking for max_fail_percentage 13731 1727203835.11157: done checking for max_fail_percentage 13731 1727203835.11158: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.11159: done checking to see if all hosts have failed 13731 1727203835.11159: getting the remaining hosts for this loop 13731 1727203835.11164: done getting the remaining hosts for this loop 13731 1727203835.11168: getting the next task for host managed-node3 13731 1727203835.11178: done getting next task for host managed-node3 13731 1727203835.11183: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13731 1727203835.11189: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.11201: getting variables 13731 1727203835.11203: in VariableManager get_vars() 13731 1727203835.11240: Calling all_inventory to load vars for managed-node3 13731 1727203835.11243: Calling groups_inventory to load vars for managed-node3 13731 1727203835.11246: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.11255: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.11259: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.11264: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.13125: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.15910: done with get_vars() 13731 1727203835.15945: done getting variables 13731 1727203835.16243: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:11 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.093) 0:00:13.393 ***** 13731 1727203835.16289: entering _queue_task() for managed-node3/fail 13731 1727203835.16291: Creating lock for fail 13731 1727203835.16644: worker is 1 (out of 1 available) 13731 1727203835.16658: exiting _queue_task() for managed-node3/fail 13731 1727203835.16671: done queuing things up, now waiting for results queue to drain 13731 1727203835.16672: waiting for pending results... 13731 1727203835.16998: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13731 1727203835.17156: in run() - task 028d2410-947f-82dc-c122-000000000279 13731 1727203835.17182: variable 'ansible_search_path' from source: unknown 13731 1727203835.17193: variable 'ansible_search_path' from source: unknown 13731 1727203835.17242: calling self._execute() 13731 1727203835.17354: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.17380: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.17396: variable 'omit' from source: magic vars 13731 1727203835.17797: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.17903: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.17944: variable 'network_state' from source: role '' defaults 13731 1727203835.17959: Evaluated conditional (network_state != {}): False 13731 1727203835.17970: when evaluation is False, skipping this task 13731 1727203835.17979: _execute() done 13731 1727203835.17986: dumping result to json 13731 1727203835.17993: done dumping result, returning 13731 1727203835.18002: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider [028d2410-947f-82dc-c122-000000000279] 13731 1727203835.18019: sending task result for task 028d2410-947f-82dc-c122-000000000279 13731 1727203835.18290: done sending task result for task 028d2410-947f-82dc-c122-000000000279 13731 1727203835.18294: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203835.18346: no more pending results, returning what we have 13731 1727203835.18351: results queue empty 13731 1727203835.18352: checking for any_errors_fatal 13731 1727203835.18358: done checking for any_errors_fatal 13731 1727203835.18359: checking for max_fail_percentage 13731 1727203835.18361: done checking for max_fail_percentage 13731 1727203835.18364: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.18365: done checking to see if all hosts have failed 13731 1727203835.18366: getting the remaining hosts for this loop 13731 1727203835.18368: done getting the remaining hosts for this loop 13731 1727203835.18371: getting the next task for host managed-node3 13731 1727203835.18381: done getting next task for host managed-node3 13731 1727203835.18385: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13731 1727203835.18391: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.18407: getting variables 13731 1727203835.18409: in VariableManager get_vars() 13731 1727203835.18446: Calling all_inventory to load vars for managed-node3 13731 1727203835.18449: Calling groups_inventory to load vars for managed-node3 13731 1727203835.18452: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.18466: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.18469: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.18472: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.19952: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.21521: done with get_vars() 13731 1727203835.21550: done getting variables 13731 1727203835.21612: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:18 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.053) 0:00:13.447 ***** 13731 1727203835.21644: entering _queue_task() for managed-node3/fail 13731 1727203835.21969: worker is 1 (out of 1 available) 13731 1727203835.22085: exiting _queue_task() for managed-node3/fail 13731 1727203835.22096: done queuing things up, now waiting for results queue to drain 13731 1727203835.22098: waiting for pending results... 13731 1727203835.22269: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13731 1727203835.22403: in run() - task 028d2410-947f-82dc-c122-00000000027a 13731 1727203835.22581: variable 'ansible_search_path' from source: unknown 13731 1727203835.22585: variable 'ansible_search_path' from source: unknown 13731 1727203835.22587: calling self._execute() 13731 1727203835.22590: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.22592: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.22595: variable 'omit' from source: magic vars 13731 1727203835.22953: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.22973: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.23104: variable 'network_state' from source: role '' defaults 13731 1727203835.23120: Evaluated conditional (network_state != {}): False 13731 1727203835.23128: when evaluation is False, skipping this task 13731 1727203835.23138: _execute() done 13731 1727203835.23147: dumping result to json 13731 1727203835.23153: done dumping result, returning 13731 1727203835.23164: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 [028d2410-947f-82dc-c122-00000000027a] 13731 1727203835.23174: sending task result for task 028d2410-947f-82dc-c122-00000000027a skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203835.23442: no more pending results, returning what we have 13731 1727203835.23446: results queue empty 13731 1727203835.23446: checking for any_errors_fatal 13731 1727203835.23459: done checking for any_errors_fatal 13731 1727203835.23460: checking for max_fail_percentage 13731 1727203835.23463: done checking for max_fail_percentage 13731 1727203835.23464: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.23465: done checking to see if all hosts have failed 13731 1727203835.23466: getting the remaining hosts for this loop 13731 1727203835.23467: done getting the remaining hosts for this loop 13731 1727203835.23471: getting the next task for host managed-node3 13731 1727203835.23480: done getting next task for host managed-node3 13731 1727203835.23485: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13731 1727203835.23490: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.23506: getting variables 13731 1727203835.23508: in VariableManager get_vars() 13731 1727203835.23545: Calling all_inventory to load vars for managed-node3 13731 1727203835.23548: Calling groups_inventory to load vars for managed-node3 13731 1727203835.23550: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.23565: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.23568: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.23572: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.23779: done sending task result for task 028d2410-947f-82dc-c122-00000000027a 13731 1727203835.23782: WORKER PROCESS EXITING 13731 1727203835.25303: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.26898: done with get_vars() 13731 1727203835.26925: done getting variables 13731 1727203835.26991: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:25 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.053) 0:00:13.501 ***** 13731 1727203835.27033: entering _queue_task() for managed-node3/fail 13731 1727203835.27395: worker is 1 (out of 1 available) 13731 1727203835.27407: exiting _queue_task() for managed-node3/fail 13731 1727203835.27418: done queuing things up, now waiting for results queue to drain 13731 1727203835.27420: waiting for pending results... 13731 1727203835.27716: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13731 1727203835.27998: in run() - task 028d2410-947f-82dc-c122-00000000027b 13731 1727203835.28169: variable 'ansible_search_path' from source: unknown 13731 1727203835.28181: variable 'ansible_search_path' from source: unknown 13731 1727203835.28216: calling self._execute() 13731 1727203835.28479: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.28494: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.28510: variable 'omit' from source: magic vars 13731 1727203835.29352: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.29355: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.29458: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203835.32023: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203835.32156: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203835.32312: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203835.32347: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203835.32407: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203835.32571: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.32868: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.32872: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.32978: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.32982: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.33222: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.33286: Evaluated conditional (ansible_distribution_major_version | int > 9): True 13731 1727203835.33440: variable 'ansible_distribution' from source: facts 13731 1727203835.33451: variable '__network_rh_distros' from source: role '' defaults 13731 1727203835.33469: Evaluated conditional (ansible_distribution in __network_rh_distros): True 13731 1727203835.33757: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.33793: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.33823: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.33877: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.33898: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.33980: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.34015: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.34045: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.34108: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.34131: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.34186: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.34227: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.34257: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.34316: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.34337: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.34693: variable 'network_connections' from source: include params 13731 1727203835.34782: variable 'controller_profile' from source: play vars 13731 1727203835.34788: variable 'controller_profile' from source: play vars 13731 1727203835.34804: variable 'controller_device' from source: play vars 13731 1727203835.34874: variable 'controller_device' from source: play vars 13731 1727203835.34895: variable 'port1_profile' from source: play vars 13731 1727203835.34959: variable 'port1_profile' from source: play vars 13731 1727203835.34978: variable 'dhcp_interface1' from source: play vars 13731 1727203835.35044: variable 'dhcp_interface1' from source: play vars 13731 1727203835.35057: variable 'controller_profile' from source: play vars 13731 1727203835.35122: variable 'controller_profile' from source: play vars 13731 1727203835.35139: variable 'port2_profile' from source: play vars 13731 1727203835.35205: variable 'port2_profile' from source: play vars 13731 1727203835.35245: variable 'dhcp_interface2' from source: play vars 13731 1727203835.35286: variable 'dhcp_interface2' from source: play vars 13731 1727203835.35299: variable 'controller_profile' from source: play vars 13731 1727203835.35381: variable 'controller_profile' from source: play vars 13731 1727203835.35464: variable 'network_state' from source: role '' defaults 13731 1727203835.35467: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203835.35773: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203835.35823: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203835.35860: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203835.35910: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203835.35980: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203835.36009: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203835.36041: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.36081: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203835.36129: Evaluated conditional (network_connections | selectattr("type", "defined") | selectattr("type", "match", "^team$") | list | length > 0 or network_state.get("interfaces", []) | selectattr("type", "defined") | selectattr("type", "match", "^team$") | list | length > 0): False 13731 1727203835.36136: when evaluation is False, skipping this task 13731 1727203835.36180: _execute() done 13731 1727203835.36183: dumping result to json 13731 1727203835.36185: done dumping result, returning 13731 1727203835.36187: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later [028d2410-947f-82dc-c122-00000000027b] 13731 1727203835.36190: sending task result for task 028d2410-947f-82dc-c122-00000000027b 13731 1727203835.36484: done sending task result for task 028d2410-947f-82dc-c122-00000000027b 13731 1727203835.36487: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_connections | selectattr(\"type\", \"defined\") | selectattr(\"type\", \"match\", \"^team$\") | list | length > 0 or network_state.get(\"interfaces\", []) | selectattr(\"type\", \"defined\") | selectattr(\"type\", \"match\", \"^team$\") | list | length > 0", "skip_reason": "Conditional result was False" } 13731 1727203835.36547: no more pending results, returning what we have 13731 1727203835.36551: results queue empty 13731 1727203835.36552: checking for any_errors_fatal 13731 1727203835.36558: done checking for any_errors_fatal 13731 1727203835.36558: checking for max_fail_percentage 13731 1727203835.36560: done checking for max_fail_percentage 13731 1727203835.36561: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.36564: done checking to see if all hosts have failed 13731 1727203835.36565: getting the remaining hosts for this loop 13731 1727203835.36567: done getting the remaining hosts for this loop 13731 1727203835.36570: getting the next task for host managed-node3 13731 1727203835.36580: done getting next task for host managed-node3 13731 1727203835.36584: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13731 1727203835.36590: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.36604: getting variables 13731 1727203835.36607: in VariableManager get_vars() 13731 1727203835.36645: Calling all_inventory to load vars for managed-node3 13731 1727203835.36649: Calling groups_inventory to load vars for managed-node3 13731 1727203835.36651: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.36664: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.36667: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.36670: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.38150: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.40117: done with get_vars() 13731 1727203835.40152: done getting variables 13731 1727203835.40297: Loading ActionModule 'dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:36 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.132) 0:00:13.634 ***** 13731 1727203835.40332: entering _queue_task() for managed-node3/dnf 13731 1727203835.40899: worker is 1 (out of 1 available) 13731 1727203835.40909: exiting _queue_task() for managed-node3/dnf 13731 1727203835.40919: done queuing things up, now waiting for results queue to drain 13731 1727203835.40920: waiting for pending results... 13731 1727203835.41016: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13731 1727203835.41179: in run() - task 028d2410-947f-82dc-c122-00000000027c 13731 1727203835.41202: variable 'ansible_search_path' from source: unknown 13731 1727203835.41211: variable 'ansible_search_path' from source: unknown 13731 1727203835.41252: calling self._execute() 13731 1727203835.41369: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.41389: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.41581: variable 'omit' from source: magic vars 13731 1727203835.41834: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.41866: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.42095: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203835.45481: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203835.45485: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203835.45533: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203835.45590: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203835.45717: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203835.45813: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.45852: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.45905: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.46003: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.46027: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.46183: variable 'ansible_distribution' from source: facts 13731 1727203835.46194: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.46224: Evaluated conditional (ansible_distribution == 'Fedora' or ansible_distribution_major_version | int > 7): True 13731 1727203835.46353: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203835.46498: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.46526: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.46554: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.46604: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.46624: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.46670: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.46706: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.46783: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.46786: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.46792: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.46838: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.46869: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.46901: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.46947: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.46968: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.47128: variable 'network_connections' from source: include params 13731 1727203835.47148: variable 'controller_profile' from source: play vars 13731 1727203835.47249: variable 'controller_profile' from source: play vars 13731 1727203835.47258: variable 'controller_device' from source: play vars 13731 1727203835.47353: variable 'controller_device' from source: play vars 13731 1727203835.47356: variable 'port1_profile' from source: play vars 13731 1727203835.47418: variable 'port1_profile' from source: play vars 13731 1727203835.47429: variable 'dhcp_interface1' from source: play vars 13731 1727203835.47569: variable 'dhcp_interface1' from source: play vars 13731 1727203835.47573: variable 'controller_profile' from source: play vars 13731 1727203835.47581: variable 'controller_profile' from source: play vars 13731 1727203835.47593: variable 'port2_profile' from source: play vars 13731 1727203835.47665: variable 'port2_profile' from source: play vars 13731 1727203835.47686: variable 'dhcp_interface2' from source: play vars 13731 1727203835.47785: variable 'dhcp_interface2' from source: play vars 13731 1727203835.47789: variable 'controller_profile' from source: play vars 13731 1727203835.47847: variable 'controller_profile' from source: play vars 13731 1727203835.47937: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203835.48134: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203835.48181: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203835.48217: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203835.48253: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203835.48305: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203835.48355: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203835.48453: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.48457: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203835.48487: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203835.48777: variable 'network_connections' from source: include params 13731 1727203835.48790: variable 'controller_profile' from source: play vars 13731 1727203835.48835: variable 'controller_profile' from source: play vars 13731 1727203835.48841: variable 'controller_device' from source: play vars 13731 1727203835.48886: variable 'controller_device' from source: play vars 13731 1727203835.48898: variable 'port1_profile' from source: play vars 13731 1727203835.48939: variable 'port1_profile' from source: play vars 13731 1727203835.48945: variable 'dhcp_interface1' from source: play vars 13731 1727203835.48995: variable 'dhcp_interface1' from source: play vars 13731 1727203835.49002: variable 'controller_profile' from source: play vars 13731 1727203835.49044: variable 'controller_profile' from source: play vars 13731 1727203835.49050: variable 'port2_profile' from source: play vars 13731 1727203835.49095: variable 'port2_profile' from source: play vars 13731 1727203835.49101: variable 'dhcp_interface2' from source: play vars 13731 1727203835.49144: variable 'dhcp_interface2' from source: play vars 13731 1727203835.49149: variable 'controller_profile' from source: play vars 13731 1727203835.49194: variable 'controller_profile' from source: play vars 13731 1727203835.49221: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203835.49224: when evaluation is False, skipping this task 13731 1727203835.49226: _execute() done 13731 1727203835.49229: dumping result to json 13731 1727203835.49231: done dumping result, returning 13731 1727203835.49238: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces [028d2410-947f-82dc-c122-00000000027c] 13731 1727203835.49247: sending task result for task 028d2410-947f-82dc-c122-00000000027c 13731 1727203835.49333: done sending task result for task 028d2410-947f-82dc-c122-00000000027c 13731 1727203835.49336: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203835.49391: no more pending results, returning what we have 13731 1727203835.49395: results queue empty 13731 1727203835.49395: checking for any_errors_fatal 13731 1727203835.49402: done checking for any_errors_fatal 13731 1727203835.49403: checking for max_fail_percentage 13731 1727203835.49405: done checking for max_fail_percentage 13731 1727203835.49405: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.49406: done checking to see if all hosts have failed 13731 1727203835.49407: getting the remaining hosts for this loop 13731 1727203835.49408: done getting the remaining hosts for this loop 13731 1727203835.49412: getting the next task for host managed-node3 13731 1727203835.49420: done getting next task for host managed-node3 13731 1727203835.49423: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13731 1727203835.49428: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.49442: getting variables 13731 1727203835.49444: in VariableManager get_vars() 13731 1727203835.49482: Calling all_inventory to load vars for managed-node3 13731 1727203835.49485: Calling groups_inventory to load vars for managed-node3 13731 1727203835.49487: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.49497: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.49499: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.49501: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.50291: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.51620: done with get_vars() 13731 1727203835.51653: done getting variables redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13731 1727203835.51711: Loading ActionModule 'ansible_collections.ansible.builtin.plugins.action.dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:48 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.114) 0:00:13.748 ***** 13731 1727203835.51734: entering _queue_task() for managed-node3/yum 13731 1727203835.51736: Creating lock for yum 13731 1727203835.52022: worker is 1 (out of 1 available) 13731 1727203835.52036: exiting _queue_task() for managed-node3/yum 13731 1727203835.52048: done queuing things up, now waiting for results queue to drain 13731 1727203835.52050: waiting for pending results... 13731 1727203835.52232: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13731 1727203835.52321: in run() - task 028d2410-947f-82dc-c122-00000000027d 13731 1727203835.52333: variable 'ansible_search_path' from source: unknown 13731 1727203835.52337: variable 'ansible_search_path' from source: unknown 13731 1727203835.52367: calling self._execute() 13731 1727203835.52435: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.52439: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.52448: variable 'omit' from source: magic vars 13731 1727203835.52711: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.52725: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.52840: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203835.54649: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203835.54653: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203835.54658: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203835.54717: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203835.54750: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203835.54846: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.54893: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.54922: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.54966: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.54997: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.55104: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.55124: Evaluated conditional (ansible_distribution_major_version | int < 8): False 13731 1727203835.55131: when evaluation is False, skipping this task 13731 1727203835.55138: _execute() done 13731 1727203835.55145: dumping result to json 13731 1727203835.55152: done dumping result, returning 13731 1727203835.55164: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces [028d2410-947f-82dc-c122-00000000027d] 13731 1727203835.55174: sending task result for task 028d2410-947f-82dc-c122-00000000027d 13731 1727203835.55300: done sending task result for task 028d2410-947f-82dc-c122-00000000027d 13731 1727203835.55306: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version | int < 8", "skip_reason": "Conditional result was False" } 13731 1727203835.55359: no more pending results, returning what we have 13731 1727203835.55362: results queue empty 13731 1727203835.55363: checking for any_errors_fatal 13731 1727203835.55370: done checking for any_errors_fatal 13731 1727203835.55371: checking for max_fail_percentage 13731 1727203835.55373: done checking for max_fail_percentage 13731 1727203835.55374: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.55375: done checking to see if all hosts have failed 13731 1727203835.55393: getting the remaining hosts for this loop 13731 1727203835.55396: done getting the remaining hosts for this loop 13731 1727203835.55400: getting the next task for host managed-node3 13731 1727203835.55407: done getting next task for host managed-node3 13731 1727203835.55411: ^ task is: TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13731 1727203835.55417: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.55430: getting variables 13731 1727203835.55432: in VariableManager get_vars() 13731 1727203835.55468: Calling all_inventory to load vars for managed-node3 13731 1727203835.55471: Calling groups_inventory to load vars for managed-node3 13731 1727203835.55473: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.55521: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.55526: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.55530: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.56998: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.58785: done with get_vars() 13731 1727203835.58811: done getting variables 13731 1727203835.58879: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:60 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.071) 0:00:13.820 ***** 13731 1727203835.58914: entering _queue_task() for managed-node3/fail 13731 1727203835.59246: worker is 1 (out of 1 available) 13731 1727203835.59259: exiting _queue_task() for managed-node3/fail 13731 1727203835.59272: done queuing things up, now waiting for results queue to drain 13731 1727203835.59273: waiting for pending results... 13731 1727203835.59693: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13731 1727203835.59727: in run() - task 028d2410-947f-82dc-c122-00000000027e 13731 1727203835.59748: variable 'ansible_search_path' from source: unknown 13731 1727203835.59766: variable 'ansible_search_path' from source: unknown 13731 1727203835.59812: calling self._execute() 13731 1727203835.59912: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.59931: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.59947: variable 'omit' from source: magic vars 13731 1727203835.60347: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.60373: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.60681: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203835.61016: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203835.63209: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203835.63268: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203835.63296: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203835.63323: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203835.63342: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203835.63403: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.63426: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.63443: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.63474: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.63489: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.63522: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.63540: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.63556: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.63586: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.63596: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.63624: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.63642: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.63658: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.63687: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.63697: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.63810: variable 'network_connections' from source: include params 13731 1727203835.63819: variable 'controller_profile' from source: play vars 13731 1727203835.63870: variable 'controller_profile' from source: play vars 13731 1727203835.63880: variable 'controller_device' from source: play vars 13731 1727203835.63925: variable 'controller_device' from source: play vars 13731 1727203835.63936: variable 'port1_profile' from source: play vars 13731 1727203835.63981: variable 'port1_profile' from source: play vars 13731 1727203835.63987: variable 'dhcp_interface1' from source: play vars 13731 1727203835.64030: variable 'dhcp_interface1' from source: play vars 13731 1727203835.64035: variable 'controller_profile' from source: play vars 13731 1727203835.64079: variable 'controller_profile' from source: play vars 13731 1727203835.64085: variable 'port2_profile' from source: play vars 13731 1727203835.64126: variable 'port2_profile' from source: play vars 13731 1727203835.64132: variable 'dhcp_interface2' from source: play vars 13731 1727203835.64184: variable 'dhcp_interface2' from source: play vars 13731 1727203835.64381: variable 'controller_profile' from source: play vars 13731 1727203835.64384: variable 'controller_profile' from source: play vars 13731 1727203835.64387: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203835.64540: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203835.64585: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203835.64628: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203835.64663: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203835.64711: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203835.64747: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203835.64781: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.64810: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203835.64895: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203835.65182: variable 'network_connections' from source: include params 13731 1727203835.65223: variable 'controller_profile' from source: play vars 13731 1727203835.65246: variable 'controller_profile' from source: play vars 13731 1727203835.65253: variable 'controller_device' from source: play vars 13731 1727203835.65302: variable 'controller_device' from source: play vars 13731 1727203835.65312: variable 'port1_profile' from source: play vars 13731 1727203835.65352: variable 'port1_profile' from source: play vars 13731 1727203835.65358: variable 'dhcp_interface1' from source: play vars 13731 1727203835.65406: variable 'dhcp_interface1' from source: play vars 13731 1727203835.65411: variable 'controller_profile' from source: play vars 13731 1727203835.65451: variable 'controller_profile' from source: play vars 13731 1727203835.65462: variable 'port2_profile' from source: play vars 13731 1727203835.65509: variable 'port2_profile' from source: play vars 13731 1727203835.65515: variable 'dhcp_interface2' from source: play vars 13731 1727203835.65555: variable 'dhcp_interface2' from source: play vars 13731 1727203835.65560: variable 'controller_profile' from source: play vars 13731 1727203835.65607: variable 'controller_profile' from source: play vars 13731 1727203835.65631: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203835.65634: when evaluation is False, skipping this task 13731 1727203835.65637: _execute() done 13731 1727203835.65639: dumping result to json 13731 1727203835.65642: done dumping result, returning 13731 1727203835.65649: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces [028d2410-947f-82dc-c122-00000000027e] 13731 1727203835.65653: sending task result for task 028d2410-947f-82dc-c122-00000000027e 13731 1727203835.65741: done sending task result for task 028d2410-947f-82dc-c122-00000000027e 13731 1727203835.65744: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203835.65796: no more pending results, returning what we have 13731 1727203835.65800: results queue empty 13731 1727203835.65800: checking for any_errors_fatal 13731 1727203835.65806: done checking for any_errors_fatal 13731 1727203835.65806: checking for max_fail_percentage 13731 1727203835.65808: done checking for max_fail_percentage 13731 1727203835.65808: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.65809: done checking to see if all hosts have failed 13731 1727203835.65810: getting the remaining hosts for this loop 13731 1727203835.65812: done getting the remaining hosts for this loop 13731 1727203835.65815: getting the next task for host managed-node3 13731 1727203835.65822: done getting next task for host managed-node3 13731 1727203835.65825: ^ task is: TASK: fedora.linux_system_roles.network : Install packages 13731 1727203835.65830: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.65844: getting variables 13731 1727203835.65846: in VariableManager get_vars() 13731 1727203835.65884: Calling all_inventory to load vars for managed-node3 13731 1727203835.65886: Calling groups_inventory to load vars for managed-node3 13731 1727203835.65889: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.65898: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.65900: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.65903: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.66691: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.67546: done with get_vars() 13731 1727203835.67561: done getting variables 13731 1727203835.67605: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install packages] ******************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:73 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.087) 0:00:13.907 ***** 13731 1727203835.67629: entering _queue_task() for managed-node3/package 13731 1727203835.67839: worker is 1 (out of 1 available) 13731 1727203835.67852: exiting _queue_task() for managed-node3/package 13731 1727203835.67868: done queuing things up, now waiting for results queue to drain 13731 1727203835.67869: waiting for pending results... 13731 1727203835.68042: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install packages 13731 1727203835.68133: in run() - task 028d2410-947f-82dc-c122-00000000027f 13731 1727203835.68144: variable 'ansible_search_path' from source: unknown 13731 1727203835.68148: variable 'ansible_search_path' from source: unknown 13731 1727203835.68177: calling self._execute() 13731 1727203835.68242: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.68246: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.68256: variable 'omit' from source: magic vars 13731 1727203835.68515: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.68527: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.68652: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203835.68832: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203835.68867: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203835.68891: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203835.68915: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203835.68994: variable 'network_packages' from source: role '' defaults 13731 1727203835.69066: variable '__network_provider_setup' from source: role '' defaults 13731 1727203835.69069: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203835.69117: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203835.69124: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203835.69168: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203835.69280: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203835.74166: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203835.74211: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203835.74236: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203835.74267: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203835.74297: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203835.74346: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.74368: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.74387: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.74416: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.74427: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.74456: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.74473: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.74491: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.74520: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.74530: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.74670: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13731 1727203835.74747: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.74766: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.74783: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.74807: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.74818: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.74880: variable 'ansible_python' from source: facts 13731 1727203835.74893: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13731 1727203835.74951: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203835.75006: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203835.75092: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.75108: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.75124: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.75148: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.75159: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.75197: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203835.75216: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203835.75232: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.75256: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203835.75268: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203835.75362: variable 'network_connections' from source: include params 13731 1727203835.75369: variable 'controller_profile' from source: play vars 13731 1727203835.75441: variable 'controller_profile' from source: play vars 13731 1727203835.75449: variable 'controller_device' from source: play vars 13731 1727203835.75522: variable 'controller_device' from source: play vars 13731 1727203835.75533: variable 'port1_profile' from source: play vars 13731 1727203835.75605: variable 'port1_profile' from source: play vars 13731 1727203835.75613: variable 'dhcp_interface1' from source: play vars 13731 1727203835.75681: variable 'dhcp_interface1' from source: play vars 13731 1727203835.75689: variable 'controller_profile' from source: play vars 13731 1727203835.75757: variable 'controller_profile' from source: play vars 13731 1727203835.75767: variable 'port2_profile' from source: play vars 13731 1727203835.75837: variable 'port2_profile' from source: play vars 13731 1727203835.75845: variable 'dhcp_interface2' from source: play vars 13731 1727203835.75917: variable 'dhcp_interface2' from source: play vars 13731 1727203835.75926: variable 'controller_profile' from source: play vars 13731 1727203835.75992: variable 'controller_profile' from source: play vars 13731 1727203835.76280: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203835.76283: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203835.76285: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203835.76287: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203835.76289: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203835.76545: variable 'network_connections' from source: include params 13731 1727203835.76554: variable 'controller_profile' from source: play vars 13731 1727203835.76652: variable 'controller_profile' from source: play vars 13731 1727203835.76671: variable 'controller_device' from source: play vars 13731 1727203835.76791: variable 'controller_device' from source: play vars 13731 1727203835.76809: variable 'port1_profile' from source: play vars 13731 1727203835.76910: variable 'port1_profile' from source: play vars 13731 1727203835.76923: variable 'dhcp_interface1' from source: play vars 13731 1727203835.77021: variable 'dhcp_interface1' from source: play vars 13731 1727203835.77035: variable 'controller_profile' from source: play vars 13731 1727203835.77134: variable 'controller_profile' from source: play vars 13731 1727203835.77149: variable 'port2_profile' from source: play vars 13731 1727203835.77248: variable 'port2_profile' from source: play vars 13731 1727203835.77265: variable 'dhcp_interface2' from source: play vars 13731 1727203835.77372: variable 'dhcp_interface2' from source: play vars 13731 1727203835.77388: variable 'controller_profile' from source: play vars 13731 1727203835.77490: variable 'controller_profile' from source: play vars 13731 1727203835.77545: variable '__network_packages_default_wireless' from source: role '' defaults 13731 1727203835.77633: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203835.77936: variable 'network_connections' from source: include params 13731 1727203835.77945: variable 'controller_profile' from source: play vars 13731 1727203835.78013: variable 'controller_profile' from source: play vars 13731 1727203835.78027: variable 'controller_device' from source: play vars 13731 1727203835.78095: variable 'controller_device' from source: play vars 13731 1727203835.78111: variable 'port1_profile' from source: play vars 13731 1727203835.78178: variable 'port1_profile' from source: play vars 13731 1727203835.78380: variable 'dhcp_interface1' from source: play vars 13731 1727203835.78384: variable 'dhcp_interface1' from source: play vars 13731 1727203835.78386: variable 'controller_profile' from source: play vars 13731 1727203835.78388: variable 'controller_profile' from source: play vars 13731 1727203835.78390: variable 'port2_profile' from source: play vars 13731 1727203835.78401: variable 'port2_profile' from source: play vars 13731 1727203835.78412: variable 'dhcp_interface2' from source: play vars 13731 1727203835.78480: variable 'dhcp_interface2' from source: play vars 13731 1727203835.78492: variable 'controller_profile' from source: play vars 13731 1727203835.78555: variable 'controller_profile' from source: play vars 13731 1727203835.78590: variable '__network_packages_default_team' from source: role '' defaults 13731 1727203835.78671: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203835.78987: variable 'network_connections' from source: include params 13731 1727203835.78997: variable 'controller_profile' from source: play vars 13731 1727203835.79064: variable 'controller_profile' from source: play vars 13731 1727203835.79078: variable 'controller_device' from source: play vars 13731 1727203835.79142: variable 'controller_device' from source: play vars 13731 1727203835.79159: variable 'port1_profile' from source: play vars 13731 1727203835.79226: variable 'port1_profile' from source: play vars 13731 1727203835.79238: variable 'dhcp_interface1' from source: play vars 13731 1727203835.79309: variable 'dhcp_interface1' from source: play vars 13731 1727203835.79323: variable 'controller_profile' from source: play vars 13731 1727203835.79394: variable 'controller_profile' from source: play vars 13731 1727203835.79406: variable 'port2_profile' from source: play vars 13731 1727203835.79471: variable 'port2_profile' from source: play vars 13731 1727203835.79487: variable 'dhcp_interface2' from source: play vars 13731 1727203835.79550: variable 'dhcp_interface2' from source: play vars 13731 1727203835.79561: variable 'controller_profile' from source: play vars 13731 1727203835.79628: variable 'controller_profile' from source: play vars 13731 1727203835.79698: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203835.79759: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203835.79773: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203835.79836: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203835.80040: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13731 1727203835.80465: variable 'network_connections' from source: include params 13731 1727203835.80479: variable 'controller_profile' from source: play vars 13731 1727203835.80544: variable 'controller_profile' from source: play vars 13731 1727203835.80565: variable 'controller_device' from source: play vars 13731 1727203835.80632: variable 'controller_device' from source: play vars 13731 1727203835.80650: variable 'port1_profile' from source: play vars 13731 1727203835.80743: variable 'port1_profile' from source: play vars 13731 1727203835.80755: variable 'dhcp_interface1' from source: play vars 13731 1727203835.80839: variable 'dhcp_interface1' from source: play vars 13731 1727203835.80858: variable 'controller_profile' from source: play vars 13731 1727203835.80931: variable 'controller_profile' from source: play vars 13731 1727203835.80959: variable 'port2_profile' from source: play vars 13731 1727203835.81070: variable 'port2_profile' from source: play vars 13731 1727203835.81074: variable 'dhcp_interface2' from source: play vars 13731 1727203835.81110: variable 'dhcp_interface2' from source: play vars 13731 1727203835.81122: variable 'controller_profile' from source: play vars 13731 1727203835.81190: variable 'controller_profile' from source: play vars 13731 1727203835.81280: variable 'ansible_distribution' from source: facts 13731 1727203835.81285: variable '__network_rh_distros' from source: role '' defaults 13731 1727203835.81288: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.81290: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13731 1727203835.81434: variable 'ansible_distribution' from source: facts 13731 1727203835.81443: variable '__network_rh_distros' from source: role '' defaults 13731 1727203835.81454: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.81474: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13731 1727203835.81684: variable 'ansible_distribution' from source: facts 13731 1727203835.81703: variable '__network_rh_distros' from source: role '' defaults 13731 1727203835.81731: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.81853: variable 'network_provider' from source: set_fact 13731 1727203835.81856: variable 'ansible_facts' from source: unknown 13731 1727203835.82655: Evaluated conditional (not network_packages is subset(ansible_facts.packages.keys())): False 13731 1727203835.82664: when evaluation is False, skipping this task 13731 1727203835.82671: _execute() done 13731 1727203835.82677: dumping result to json 13731 1727203835.82684: done dumping result, returning 13731 1727203835.82693: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install packages [028d2410-947f-82dc-c122-00000000027f] 13731 1727203835.82701: sending task result for task 028d2410-947f-82dc-c122-00000000027f skipping: [managed-node3] => { "changed": false, "false_condition": "not network_packages is subset(ansible_facts.packages.keys())", "skip_reason": "Conditional result was False" } 13731 1727203835.83025: no more pending results, returning what we have 13731 1727203835.83029: results queue empty 13731 1727203835.83030: checking for any_errors_fatal 13731 1727203835.83034: done checking for any_errors_fatal 13731 1727203835.83035: checking for max_fail_percentage 13731 1727203835.83036: done checking for max_fail_percentage 13731 1727203835.83037: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.83038: done checking to see if all hosts have failed 13731 1727203835.83039: getting the remaining hosts for this loop 13731 1727203835.83041: done getting the remaining hosts for this loop 13731 1727203835.83044: getting the next task for host managed-node3 13731 1727203835.83052: done getting next task for host managed-node3 13731 1727203835.83057: ^ task is: TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13731 1727203835.83064: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.83080: getting variables 13731 1727203835.83081: in VariableManager get_vars() 13731 1727203835.83116: Calling all_inventory to load vars for managed-node3 13731 1727203835.83118: Calling groups_inventory to load vars for managed-node3 13731 1727203835.83121: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.83130: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.83132: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.83135: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.83709: done sending task result for task 028d2410-947f-82dc-c122-00000000027f 13731 1727203835.83713: WORKER PROCESS EXITING 13731 1727203835.88440: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.90007: done with get_vars() 13731 1727203835.90030: done getting variables 13731 1727203835.90082: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:85 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.224) 0:00:14.132 ***** 13731 1727203835.90112: entering _queue_task() for managed-node3/package 13731 1727203835.90432: worker is 1 (out of 1 available) 13731 1727203835.90445: exiting _queue_task() for managed-node3/package 13731 1727203835.90458: done queuing things up, now waiting for results queue to drain 13731 1727203835.90460: waiting for pending results... 13731 1727203835.90737: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13731 1727203835.90889: in run() - task 028d2410-947f-82dc-c122-000000000280 13731 1727203835.90915: variable 'ansible_search_path' from source: unknown 13731 1727203835.90924: variable 'ansible_search_path' from source: unknown 13731 1727203835.90967: calling self._execute() 13731 1727203835.91067: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.91082: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.91098: variable 'omit' from source: magic vars 13731 1727203835.91472: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.91489: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.91616: variable 'network_state' from source: role '' defaults 13731 1727203835.91633: Evaluated conditional (network_state != {}): False 13731 1727203835.91641: when evaluation is False, skipping this task 13731 1727203835.91648: _execute() done 13731 1727203835.91657: dumping result to json 13731 1727203835.91671: done dumping result, returning 13731 1727203835.91684: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable [028d2410-947f-82dc-c122-000000000280] 13731 1727203835.91695: sending task result for task 028d2410-947f-82dc-c122-000000000280 skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203835.91866: no more pending results, returning what we have 13731 1727203835.91872: results queue empty 13731 1727203835.91873: checking for any_errors_fatal 13731 1727203835.91885: done checking for any_errors_fatal 13731 1727203835.91886: checking for max_fail_percentage 13731 1727203835.91888: done checking for max_fail_percentage 13731 1727203835.91889: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.91890: done checking to see if all hosts have failed 13731 1727203835.91890: getting the remaining hosts for this loop 13731 1727203835.91892: done getting the remaining hosts for this loop 13731 1727203835.91896: getting the next task for host managed-node3 13731 1727203835.91905: done getting next task for host managed-node3 13731 1727203835.91910: ^ task is: TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13731 1727203835.91916: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.91933: getting variables 13731 1727203835.91934: in VariableManager get_vars() 13731 1727203835.92278: Calling all_inventory to load vars for managed-node3 13731 1727203835.92282: Calling groups_inventory to load vars for managed-node3 13731 1727203835.92285: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.92295: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.92298: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.92301: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.92988: done sending task result for task 028d2410-947f-82dc-c122-000000000280 13731 1727203835.92992: WORKER PROCESS EXITING 13731 1727203835.93789: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203835.95405: done with get_vars() 13731 1727203835.95425: done getting variables 13731 1727203835.95487: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:96 Tuesday 24 September 2024 14:50:35 -0400 (0:00:00.054) 0:00:14.186 ***** 13731 1727203835.95520: entering _queue_task() for managed-node3/package 13731 1727203835.95815: worker is 1 (out of 1 available) 13731 1727203835.95827: exiting _queue_task() for managed-node3/package 13731 1727203835.95840: done queuing things up, now waiting for results queue to drain 13731 1727203835.95841: waiting for pending results... 13731 1727203835.96121: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13731 1727203835.96280: in run() - task 028d2410-947f-82dc-c122-000000000281 13731 1727203835.96305: variable 'ansible_search_path' from source: unknown 13731 1727203835.96313: variable 'ansible_search_path' from source: unknown 13731 1727203835.96353: calling self._execute() 13731 1727203835.96448: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203835.96464: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203835.96483: variable 'omit' from source: magic vars 13731 1727203835.96858: variable 'ansible_distribution_major_version' from source: facts 13731 1727203835.96878: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203835.97000: variable 'network_state' from source: role '' defaults 13731 1727203835.97015: Evaluated conditional (network_state != {}): False 13731 1727203835.97021: when evaluation is False, skipping this task 13731 1727203835.97027: _execute() done 13731 1727203835.97033: dumping result to json 13731 1727203835.97039: done dumping result, returning 13731 1727203835.97049: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable [028d2410-947f-82dc-c122-000000000281] 13731 1727203835.97484: sending task result for task 028d2410-947f-82dc-c122-000000000281 13731 1727203835.97554: done sending task result for task 028d2410-947f-82dc-c122-000000000281 13731 1727203835.97557: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203835.97602: no more pending results, returning what we have 13731 1727203835.97606: results queue empty 13731 1727203835.97607: checking for any_errors_fatal 13731 1727203835.97614: done checking for any_errors_fatal 13731 1727203835.97615: checking for max_fail_percentage 13731 1727203835.97616: done checking for max_fail_percentage 13731 1727203835.97617: checking to see if all hosts have failed and the running result is not ok 13731 1727203835.97617: done checking to see if all hosts have failed 13731 1727203835.97618: getting the remaining hosts for this loop 13731 1727203835.97620: done getting the remaining hosts for this loop 13731 1727203835.97623: getting the next task for host managed-node3 13731 1727203835.97631: done getting next task for host managed-node3 13731 1727203835.97635: ^ task is: TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13731 1727203835.97640: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203835.97655: getting variables 13731 1727203835.97656: in VariableManager get_vars() 13731 1727203835.97693: Calling all_inventory to load vars for managed-node3 13731 1727203835.97697: Calling groups_inventory to load vars for managed-node3 13731 1727203835.97699: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203835.97709: Calling all_plugins_play to load vars for managed-node3 13731 1727203835.97712: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203835.97715: Calling groups_plugins_play to load vars for managed-node3 13731 1727203835.99407: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203836.00903: done with get_vars() 13731 1727203836.00925: done getting variables 13731 1727203836.01028: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:109 Tuesday 24 September 2024 14:50:36 -0400 (0:00:00.055) 0:00:14.241 ***** 13731 1727203836.01062: entering _queue_task() for managed-node3/service 13731 1727203836.01064: Creating lock for service 13731 1727203836.01374: worker is 1 (out of 1 available) 13731 1727203836.01388: exiting _queue_task() for managed-node3/service 13731 1727203836.01400: done queuing things up, now waiting for results queue to drain 13731 1727203836.01401: waiting for pending results... 13731 1727203836.01797: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13731 1727203836.01805: in run() - task 028d2410-947f-82dc-c122-000000000282 13731 1727203836.01823: variable 'ansible_search_path' from source: unknown 13731 1727203836.01829: variable 'ansible_search_path' from source: unknown 13731 1727203836.01866: calling self._execute() 13731 1727203836.01956: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203836.01970: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203836.01989: variable 'omit' from source: magic vars 13731 1727203836.02358: variable 'ansible_distribution_major_version' from source: facts 13731 1727203836.02374: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203836.02500: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203836.02699: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203836.04863: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203836.04948: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203836.04991: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203836.05032: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203836.05067: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203836.05148: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203836.05264: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203836.05267: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.05270: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203836.05272: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203836.05322: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203836.05349: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203836.05383: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.05426: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203836.05443: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203836.05493: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203836.05519: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203836.05545: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.05593: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203836.05612: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203836.05793: variable 'network_connections' from source: include params 13731 1727203836.05813: variable 'controller_profile' from source: play vars 13731 1727203836.05914: variable 'controller_profile' from source: play vars 13731 1727203836.05917: variable 'controller_device' from source: play vars 13731 1727203836.05963: variable 'controller_device' from source: play vars 13731 1727203836.05984: variable 'port1_profile' from source: play vars 13731 1727203836.06050: variable 'port1_profile' from source: play vars 13731 1727203836.06063: variable 'dhcp_interface1' from source: play vars 13731 1727203836.06132: variable 'dhcp_interface1' from source: play vars 13731 1727203836.06180: variable 'controller_profile' from source: play vars 13731 1727203836.06203: variable 'controller_profile' from source: play vars 13731 1727203836.06214: variable 'port2_profile' from source: play vars 13731 1727203836.06281: variable 'port2_profile' from source: play vars 13731 1727203836.06293: variable 'dhcp_interface2' from source: play vars 13731 1727203836.06356: variable 'dhcp_interface2' from source: play vars 13731 1727203836.06367: variable 'controller_profile' from source: play vars 13731 1727203836.06428: variable 'controller_profile' from source: play vars 13731 1727203836.06567: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203836.06702: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203836.06741: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203836.06774: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203836.06811: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203836.06856: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203836.06891: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203836.06980: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.06984: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203836.07004: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203836.07278: variable 'network_connections' from source: include params 13731 1727203836.07582: variable 'controller_profile' from source: play vars 13731 1727203836.07586: variable 'controller_profile' from source: play vars 13731 1727203836.07588: variable 'controller_device' from source: play vars 13731 1727203836.07620: variable 'controller_device' from source: play vars 13731 1727203836.07638: variable 'port1_profile' from source: play vars 13731 1727203836.07701: variable 'port1_profile' from source: play vars 13731 1727203836.07880: variable 'dhcp_interface1' from source: play vars 13731 1727203836.07883: variable 'dhcp_interface1' from source: play vars 13731 1727203836.07885: variable 'controller_profile' from source: play vars 13731 1727203836.07945: variable 'controller_profile' from source: play vars 13731 1727203836.08026: variable 'port2_profile' from source: play vars 13731 1727203836.08089: variable 'port2_profile' from source: play vars 13731 1727203836.08135: variable 'dhcp_interface2' from source: play vars 13731 1727203836.08296: variable 'dhcp_interface2' from source: play vars 13731 1727203836.08308: variable 'controller_profile' from source: play vars 13731 1727203836.08414: variable 'controller_profile' from source: play vars 13731 1727203836.08487: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203836.08565: when evaluation is False, skipping this task 13731 1727203836.08574: _execute() done 13731 1727203836.08585: dumping result to json 13731 1727203836.08592: done dumping result, returning 13731 1727203836.08604: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000282] 13731 1727203836.08614: sending task result for task 028d2410-947f-82dc-c122-000000000282 skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203836.08927: no more pending results, returning what we have 13731 1727203836.08931: results queue empty 13731 1727203836.08932: checking for any_errors_fatal 13731 1727203836.08940: done checking for any_errors_fatal 13731 1727203836.08941: checking for max_fail_percentage 13731 1727203836.08942: done checking for max_fail_percentage 13731 1727203836.08943: checking to see if all hosts have failed and the running result is not ok 13731 1727203836.08944: done checking to see if all hosts have failed 13731 1727203836.08945: getting the remaining hosts for this loop 13731 1727203836.08947: done getting the remaining hosts for this loop 13731 1727203836.08950: getting the next task for host managed-node3 13731 1727203836.08959: done getting next task for host managed-node3 13731 1727203836.08963: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13731 1727203836.08969: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203836.08986: getting variables 13731 1727203836.08989: in VariableManager get_vars() 13731 1727203836.09030: Calling all_inventory to load vars for managed-node3 13731 1727203836.09033: Calling groups_inventory to load vars for managed-node3 13731 1727203836.09035: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203836.09046: Calling all_plugins_play to load vars for managed-node3 13731 1727203836.09049: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203836.09052: Calling groups_plugins_play to load vars for managed-node3 13731 1727203836.09983: done sending task result for task 028d2410-947f-82dc-c122-000000000282 13731 1727203836.09987: WORKER PROCESS EXITING 13731 1727203836.11290: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203836.12953: done with get_vars() 13731 1727203836.12974: done getting variables 13731 1727203836.13030: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start NetworkManager] ***** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 Tuesday 24 September 2024 14:50:36 -0400 (0:00:00.119) 0:00:14.361 ***** 13731 1727203836.13060: entering _queue_task() for managed-node3/service 13731 1727203836.13362: worker is 1 (out of 1 available) 13731 1727203836.13578: exiting _queue_task() for managed-node3/service 13731 1727203836.13588: done queuing things up, now waiting for results queue to drain 13731 1727203836.13590: waiting for pending results... 13731 1727203836.13660: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13731 1727203836.13811: in run() - task 028d2410-947f-82dc-c122-000000000283 13731 1727203836.13835: variable 'ansible_search_path' from source: unknown 13731 1727203836.13928: variable 'ansible_search_path' from source: unknown 13731 1727203836.13932: calling self._execute() 13731 1727203836.13979: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203836.13993: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203836.14008: variable 'omit' from source: magic vars 13731 1727203836.14393: variable 'ansible_distribution_major_version' from source: facts 13731 1727203836.14412: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203836.14580: variable 'network_provider' from source: set_fact 13731 1727203836.14595: variable 'network_state' from source: role '' defaults 13731 1727203836.14611: Evaluated conditional (network_provider == "nm" or network_state != {}): True 13731 1727203836.14622: variable 'omit' from source: magic vars 13731 1727203836.14684: variable 'omit' from source: magic vars 13731 1727203836.14719: variable 'network_service_name' from source: role '' defaults 13731 1727203836.14788: variable 'network_service_name' from source: role '' defaults 13731 1727203836.14891: variable '__network_provider_setup' from source: role '' defaults 13731 1727203836.14912: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203836.14966: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203836.15021: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203836.15049: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203836.15274: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203836.17380: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203836.17462: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203836.17505: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203836.17582: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203836.17585: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203836.17661: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203836.17699: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203836.17735: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.17781: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203836.17842: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203836.17846: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203836.17874: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203836.17906: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.17951: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203836.17969: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203836.18177: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13731 1727203836.18582: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203836.18586: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203836.18588: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.18590: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203836.18681: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203836.18803: variable 'ansible_python' from source: facts 13731 1727203836.18825: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13731 1727203836.18937: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203836.19282: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203836.19358: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203836.19526: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203836.19553: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.19597: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203836.19618: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203836.19660: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203836.19741: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203836.19849: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.19892: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203836.19949: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203836.20266: variable 'network_connections' from source: include params 13731 1727203836.20280: variable 'controller_profile' from source: play vars 13731 1727203836.20355: variable 'controller_profile' from source: play vars 13731 1727203836.20583: variable 'controller_device' from source: play vars 13731 1727203836.20586: variable 'controller_device' from source: play vars 13731 1727203836.20588: variable 'port1_profile' from source: play vars 13731 1727203836.20751: variable 'port1_profile' from source: play vars 13731 1727203836.20814: variable 'dhcp_interface1' from source: play vars 13731 1727203836.20994: variable 'dhcp_interface1' from source: play vars 13731 1727203836.21011: variable 'controller_profile' from source: play vars 13731 1727203836.21134: variable 'controller_profile' from source: play vars 13731 1727203836.21342: variable 'port2_profile' from source: play vars 13731 1727203836.21345: variable 'port2_profile' from source: play vars 13731 1727203836.21348: variable 'dhcp_interface2' from source: play vars 13731 1727203836.21435: variable 'dhcp_interface2' from source: play vars 13731 1727203836.21669: variable 'controller_profile' from source: play vars 13731 1727203836.21672: variable 'controller_profile' from source: play vars 13731 1727203836.21855: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203836.22295: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203836.22434: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203836.22480: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203836.22546: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203836.22730: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203836.22764: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203836.22868: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203836.22908: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203836.22979: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203836.23260: variable 'network_connections' from source: include params 13731 1727203836.23278: variable 'controller_profile' from source: play vars 13731 1727203836.23351: variable 'controller_profile' from source: play vars 13731 1727203836.23366: variable 'controller_device' from source: play vars 13731 1727203836.23444: variable 'controller_device' from source: play vars 13731 1727203836.23465: variable 'port1_profile' from source: play vars 13731 1727203836.23542: variable 'port1_profile' from source: play vars 13731 1727203836.23556: variable 'dhcp_interface1' from source: play vars 13731 1727203836.23634: variable 'dhcp_interface1' from source: play vars 13731 1727203836.23648: variable 'controller_profile' from source: play vars 13731 1727203836.23725: variable 'controller_profile' from source: play vars 13731 1727203836.23742: variable 'port2_profile' from source: play vars 13731 1727203836.23820: variable 'port2_profile' from source: play vars 13731 1727203836.23882: variable 'dhcp_interface2' from source: play vars 13731 1727203836.23917: variable 'dhcp_interface2' from source: play vars 13731 1727203836.23932: variable 'controller_profile' from source: play vars 13731 1727203836.24007: variable 'controller_profile' from source: play vars 13731 1727203836.24065: variable '__network_packages_default_wireless' from source: role '' defaults 13731 1727203836.24150: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203836.24447: variable 'network_connections' from source: include params 13731 1727203836.24461: variable 'controller_profile' from source: play vars 13731 1727203836.24571: variable 'controller_profile' from source: play vars 13731 1727203836.24574: variable 'controller_device' from source: play vars 13731 1727203836.24603: variable 'controller_device' from source: play vars 13731 1727203836.24617: variable 'port1_profile' from source: play vars 13731 1727203836.24685: variable 'port1_profile' from source: play vars 13731 1727203836.24699: variable 'dhcp_interface1' from source: play vars 13731 1727203836.24767: variable 'dhcp_interface1' from source: play vars 13731 1727203836.24781: variable 'controller_profile' from source: play vars 13731 1727203836.24853: variable 'controller_profile' from source: play vars 13731 1727203836.24882: variable 'port2_profile' from source: play vars 13731 1727203836.25082: variable 'port2_profile' from source: play vars 13731 1727203836.25085: variable 'dhcp_interface2' from source: play vars 13731 1727203836.25128: variable 'dhcp_interface2' from source: play vars 13731 1727203836.25282: variable 'controller_profile' from source: play vars 13731 1727203836.25305: variable 'controller_profile' from source: play vars 13731 1727203836.25550: variable '__network_packages_default_team' from source: role '' defaults 13731 1727203836.25553: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203836.26164: variable 'network_connections' from source: include params 13731 1727203836.26174: variable 'controller_profile' from source: play vars 13731 1727203836.26265: variable 'controller_profile' from source: play vars 13731 1727203836.26481: variable 'controller_device' from source: play vars 13731 1727203836.26496: variable 'controller_device' from source: play vars 13731 1727203836.26511: variable 'port1_profile' from source: play vars 13731 1727203836.26782: variable 'port1_profile' from source: play vars 13731 1727203836.26785: variable 'dhcp_interface1' from source: play vars 13731 1727203836.26828: variable 'dhcp_interface1' from source: play vars 13731 1727203836.26840: variable 'controller_profile' from source: play vars 13731 1727203836.27008: variable 'controller_profile' from source: play vars 13731 1727203836.27021: variable 'port2_profile' from source: play vars 13731 1727203836.27192: variable 'port2_profile' from source: play vars 13731 1727203836.27204: variable 'dhcp_interface2' from source: play vars 13731 1727203836.27274: variable 'dhcp_interface2' from source: play vars 13731 1727203836.27331: variable 'controller_profile' from source: play vars 13731 1727203836.27395: variable 'controller_profile' from source: play vars 13731 1727203836.27612: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203836.27715: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203836.27774: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203836.27839: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203836.28138: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13731 1727203836.28806: variable 'network_connections' from source: include params 13731 1727203836.28817: variable 'controller_profile' from source: play vars 13731 1727203836.28911: variable 'controller_profile' from source: play vars 13731 1727203836.28971: variable 'controller_device' from source: play vars 13731 1727203836.29119: variable 'controller_device' from source: play vars 13731 1727203836.29143: variable 'port1_profile' from source: play vars 13731 1727203836.29307: variable 'port1_profile' from source: play vars 13731 1727203836.29348: variable 'dhcp_interface1' from source: play vars 13731 1727203836.29834: variable 'dhcp_interface1' from source: play vars 13731 1727203836.29837: variable 'controller_profile' from source: play vars 13731 1727203836.29839: variable 'controller_profile' from source: play vars 13731 1727203836.29841: variable 'port2_profile' from source: play vars 13731 1727203836.29843: variable 'port2_profile' from source: play vars 13731 1727203836.29845: variable 'dhcp_interface2' from source: play vars 13731 1727203836.29974: variable 'dhcp_interface2' from source: play vars 13731 1727203836.30063: variable 'controller_profile' from source: play vars 13731 1727203836.30381: variable 'controller_profile' from source: play vars 13731 1727203836.30384: variable 'ansible_distribution' from source: facts 13731 1727203836.30387: variable '__network_rh_distros' from source: role '' defaults 13731 1727203836.30389: variable 'ansible_distribution_major_version' from source: facts 13731 1727203836.30391: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13731 1727203836.30622: variable 'ansible_distribution' from source: facts 13731 1727203836.30631: variable '__network_rh_distros' from source: role '' defaults 13731 1727203836.30640: variable 'ansible_distribution_major_version' from source: facts 13731 1727203836.30657: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13731 1727203836.31025: variable 'ansible_distribution' from source: facts 13731 1727203836.31033: variable '__network_rh_distros' from source: role '' defaults 13731 1727203836.31042: variable 'ansible_distribution_major_version' from source: facts 13731 1727203836.31085: variable 'network_provider' from source: set_fact 13731 1727203836.31112: variable 'omit' from source: magic vars 13731 1727203836.31380: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203836.31383: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203836.31386: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203836.31388: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203836.31392: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203836.31423: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203836.31431: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203836.31438: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203836.31539: Set connection var ansible_pipelining to False 13731 1727203836.31880: Set connection var ansible_shell_type to sh 13731 1727203836.31883: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203836.31886: Set connection var ansible_connection to ssh 13731 1727203836.31888: Set connection var ansible_shell_executable to /bin/sh 13731 1727203836.31891: Set connection var ansible_timeout to 10 13731 1727203836.31894: variable 'ansible_shell_executable' from source: unknown 13731 1727203836.31896: variable 'ansible_connection' from source: unknown 13731 1727203836.31899: variable 'ansible_module_compression' from source: unknown 13731 1727203836.31901: variable 'ansible_shell_type' from source: unknown 13731 1727203836.31904: variable 'ansible_shell_executable' from source: unknown 13731 1727203836.31906: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203836.31909: variable 'ansible_pipelining' from source: unknown 13731 1727203836.31911: variable 'ansible_timeout' from source: unknown 13731 1727203836.31914: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203836.31996: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203836.32280: variable 'omit' from source: magic vars 13731 1727203836.32283: starting attempt loop 13731 1727203836.32285: running the handler 13731 1727203836.32290: variable 'ansible_facts' from source: unknown 13731 1727203836.33630: _low_level_execute_command(): starting 13731 1727203836.33891: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203836.35107: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203836.35298: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203836.35384: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203836.35448: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203836.35643: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203836.37605: stdout chunk (state=3): >>>/root <<< 13731 1727203836.37622: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203836.37722: stderr chunk (state=3): >>><<< 13731 1727203836.37725: stdout chunk (state=3): >>><<< 13731 1727203836.37728: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203836.37748: _low_level_execute_command(): starting 13731 1727203836.37760: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693 `" && echo ansible-tmp-1727203836.3773682-14633-173256921514693="` echo /root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693 `" ) && sleep 0' 13731 1727203836.38956: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203836.39035: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203836.39255: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203836.39259: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203836.39505: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203836.41433: stdout chunk (state=3): >>>ansible-tmp-1727203836.3773682-14633-173256921514693=/root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693 <<< 13731 1727203836.41630: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203836.41633: stdout chunk (state=3): >>><<< 13731 1727203836.41636: stderr chunk (state=3): >>><<< 13731 1727203836.41638: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203836.3773682-14633-173256921514693=/root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203836.41640: variable 'ansible_module_compression' from source: unknown 13731 1727203836.41779: ANSIBALLZ: Using generic lock for ansible.legacy.systemd 13731 1727203836.41783: ANSIBALLZ: Acquiring lock 13731 1727203836.41987: ANSIBALLZ: Lock acquired: 140078454804688 13731 1727203836.41991: ANSIBALLZ: Creating module 13731 1727203836.80091: ANSIBALLZ: Writing module into payload 13731 1727203836.80227: ANSIBALLZ: Writing module 13731 1727203836.80248: ANSIBALLZ: Renaming module 13731 1727203836.80251: ANSIBALLZ: Done creating module 13731 1727203836.80302: variable 'ansible_facts' from source: unknown 13731 1727203836.80463: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/AnsiballZ_systemd.py 13731 1727203836.80696: Sending initial data 13731 1727203836.80699: Sent initial data (156 bytes) 13731 1727203836.81272: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203836.81321: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203836.81324: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203836.81365: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203836.82994: stderr chunk (state=3): >>>debug2: Remote version: 3 <<< 13731 1727203836.83004: stderr chunk (state=3): >>>debug2: Server supports extension "posix-rename@openssh.com" revision 1 <<< 13731 1727203836.83012: stderr chunk (state=3): >>>debug2: Server supports extension "statvfs@openssh.com" revision 2 <<< 13731 1727203836.83029: stderr chunk (state=3): >>>debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203836.83066: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203836.83097: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpk8xcizli /root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/AnsiballZ_systemd.py <<< 13731 1727203836.83104: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/AnsiballZ_systemd.py" <<< 13731 1727203836.83131: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpk8xcizli" to remote "/root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/AnsiballZ_systemd.py" <<< 13731 1727203836.83135: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/AnsiballZ_systemd.py" <<< 13731 1727203836.84208: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203836.84366: stderr chunk (state=3): >>><<< 13731 1727203836.84370: stdout chunk (state=3): >>><<< 13731 1727203836.84372: done transferring module to remote 13731 1727203836.84374: _low_level_execute_command(): starting 13731 1727203836.84383: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/ /root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/AnsiballZ_systemd.py && sleep 0' 13731 1727203836.84895: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203836.84908: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203836.84957: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203836.84970: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203836.86712: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203836.86735: stderr chunk (state=3): >>><<< 13731 1727203836.86738: stdout chunk (state=3): >>><<< 13731 1727203836.86751: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203836.86754: _low_level_execute_command(): starting 13731 1727203836.86760: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/AnsiballZ_systemd.py && sleep 0' 13731 1727203836.87168: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203836.87203: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203836.87207: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203836.87211: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203836.87213: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203836.87215: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203836.87265: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203836.87268: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203836.87274: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203836.87313: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203837.16347: stdout chunk (state=3): >>> {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "RestartMode": "normal", "NotifyAccess": "none", "RestartUSec": "100ms", "RestartSteps": "0", "RestartMaxDelayUSec": "infinity", "RestartUSecNext": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "704", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "FileDescriptorStorePreserve": "restart", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainStartTimestampMonotonic": "28990148", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainHandoffTimestampMonotonic": "29005881", "ExecMainPID": "704", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "2938", "MemoryCurrent": "10240000", "MemoryPeak": "13586432", "MemorySwapCurrent": "0", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryAvailable": "3307741184", "EffectiveMemoryMax": "3702865920", "EffectiveMemoryHigh": "3702865920", "CPUUsageNSec": "565230000", "TasksCurrent": "4", "EffectiveTasksMax": "22362", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultStartupMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "StartupMemoryLow": "0", "MemoryHigh": "infinity", "StartupMemoryHigh": "infinity", "MemoryMax": "infinity", "StartupMemoryMax": "infinity", "MemorySwapMax": "infinity", "StartupMemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22362", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "MemoryPressureWatch": "auto", "MemoryPressureThresholdUSec": "200ms", "CoredumpRe<<< 13731 1727203837.16380: stdout chunk (state=3): >>>ceive": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "RootEphemeral": "no", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "SetLoginEnvironment": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "MemoryKSM": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "system.slice dbus.socket sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "cloud-init.service shutdown.target NetworkManager-wait-online.service multi-user.target network.target", "After": "dbus-br<<< 13731 1727203837.16398: stdout chunk (state=3): >>>oker.service systemd-journald.socket network-pre.target basic.target cloud-init-local.service dbus.socket system.slice sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:49:45 EDT", "StateChangeTimestampMonotonic": "362725592", "InactiveExitTimestamp": "Tue 2024-09-24 14:44:11 EDT", "InactiveExitTimestampMonotonic": "28990654", "ActiveEnterTimestamp": "Tue 2024-09-24 14:44:12 EDT", "ActiveEnterTimestampMonotonic": "29769382", "ActiveExitTimestampMonotonic": "0", "InactiveEnterTimestampMonotonic": "0", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "SurviveFinalKillSignal": "no", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ConditionTimestampMonotonic": "28989295", "AssertTimestamp": "Tue 2024-09-24 14:44:11 EDT", "AssertTimestampMonotonic": "28989297", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "70a845f8a1964db89963090ed497f47f", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 13731 1727203837.18238: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203837.18241: stdout chunk (state=3): >>><<< 13731 1727203837.18244: stderr chunk (state=3): >>><<< 13731 1727203837.18482: _low_level_execute_command() done: rc=0, stdout= {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "RestartMode": "normal", "NotifyAccess": "none", "RestartUSec": "100ms", "RestartSteps": "0", "RestartMaxDelayUSec": "infinity", "RestartUSecNext": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "704", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "FileDescriptorStorePreserve": "restart", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainStartTimestampMonotonic": "28990148", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainHandoffTimestampMonotonic": "29005881", "ExecMainPID": "704", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "2938", "MemoryCurrent": "10240000", "MemoryPeak": "13586432", "MemorySwapCurrent": "0", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryAvailable": "3307741184", "EffectiveMemoryMax": "3702865920", "EffectiveMemoryHigh": "3702865920", "CPUUsageNSec": "565230000", "TasksCurrent": "4", "EffectiveTasksMax": "22362", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultStartupMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "StartupMemoryLow": "0", "MemoryHigh": "infinity", "StartupMemoryHigh": "infinity", "MemoryMax": "infinity", "StartupMemoryMax": "infinity", "MemorySwapMax": "infinity", "StartupMemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22362", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "MemoryPressureWatch": "auto", "MemoryPressureThresholdUSec": "200ms", "CoredumpReceive": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "RootEphemeral": "no", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "SetLoginEnvironment": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "MemoryKSM": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "system.slice dbus.socket sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "cloud-init.service shutdown.target NetworkManager-wait-online.service multi-user.target network.target", "After": "dbus-broker.service systemd-journald.socket network-pre.target basic.target cloud-init-local.service dbus.socket system.slice sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:49:45 EDT", "StateChangeTimestampMonotonic": "362725592", "InactiveExitTimestamp": "Tue 2024-09-24 14:44:11 EDT", "InactiveExitTimestampMonotonic": "28990654", "ActiveEnterTimestamp": "Tue 2024-09-24 14:44:12 EDT", "ActiveEnterTimestampMonotonic": "29769382", "ActiveExitTimestampMonotonic": "0", "InactiveEnterTimestampMonotonic": "0", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "SurviveFinalKillSignal": "no", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ConditionTimestampMonotonic": "28989295", "AssertTimestamp": "Tue 2024-09-24 14:44:11 EDT", "AssertTimestampMonotonic": "28989297", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "70a845f8a1964db89963090ed497f47f", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203837.18493: done with _execute_module (ansible.legacy.systemd, {'name': 'NetworkManager', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203837.18521: _low_level_execute_command(): starting 13731 1727203837.18531: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203836.3773682-14633-173256921514693/ > /dev/null 2>&1 && sleep 0' 13731 1727203837.19154: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203837.19170: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203837.19185: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203837.19225: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203837.19237: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203837.19283: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203837.21068: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203837.21096: stderr chunk (state=3): >>><<< 13731 1727203837.21100: stdout chunk (state=3): >>><<< 13731 1727203837.21112: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203837.21119: handler run complete 13731 1727203837.21155: attempt loop complete, returning result 13731 1727203837.21158: _execute() done 13731 1727203837.21160: dumping result to json 13731 1727203837.21177: done dumping result, returning 13731 1727203837.21186: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager [028d2410-947f-82dc-c122-000000000283] 13731 1727203837.21189: sending task result for task 028d2410-947f-82dc-c122-000000000283 13731 1727203837.21457: done sending task result for task 028d2410-947f-82dc-c122-000000000283 13731 1727203837.21461: WORKER PROCESS EXITING ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203837.21543: no more pending results, returning what we have 13731 1727203837.21547: results queue empty 13731 1727203837.21548: checking for any_errors_fatal 13731 1727203837.21555: done checking for any_errors_fatal 13731 1727203837.21556: checking for max_fail_percentage 13731 1727203837.21557: done checking for max_fail_percentage 13731 1727203837.21558: checking to see if all hosts have failed and the running result is not ok 13731 1727203837.21559: done checking to see if all hosts have failed 13731 1727203837.21559: getting the remaining hosts for this loop 13731 1727203837.21561: done getting the remaining hosts for this loop 13731 1727203837.21564: getting the next task for host managed-node3 13731 1727203837.21571: done getting next task for host managed-node3 13731 1727203837.21574: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13731 1727203837.21581: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203837.21590: getting variables 13731 1727203837.21592: in VariableManager get_vars() 13731 1727203837.21620: Calling all_inventory to load vars for managed-node3 13731 1727203837.21622: Calling groups_inventory to load vars for managed-node3 13731 1727203837.21624: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203837.21632: Calling all_plugins_play to load vars for managed-node3 13731 1727203837.21634: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203837.21636: Calling groups_plugins_play to load vars for managed-node3 13731 1727203837.23003: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203837.23887: done with get_vars() 13731 1727203837.23905: done getting variables 13731 1727203837.23950: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start wpa_supplicant] ***** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:133 Tuesday 24 September 2024 14:50:37 -0400 (0:00:01.109) 0:00:15.470 ***** 13731 1727203837.23982: entering _queue_task() for managed-node3/service 13731 1727203837.24219: worker is 1 (out of 1 available) 13731 1727203837.24233: exiting _queue_task() for managed-node3/service 13731 1727203837.24244: done queuing things up, now waiting for results queue to drain 13731 1727203837.24245: waiting for pending results... 13731 1727203837.24423: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13731 1727203837.24501: in run() - task 028d2410-947f-82dc-c122-000000000284 13731 1727203837.24513: variable 'ansible_search_path' from source: unknown 13731 1727203837.24516: variable 'ansible_search_path' from source: unknown 13731 1727203837.24544: calling self._execute() 13731 1727203837.24616: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203837.24619: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203837.24628: variable 'omit' from source: magic vars 13731 1727203837.24986: variable 'ansible_distribution_major_version' from source: facts 13731 1727203837.24995: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203837.25100: variable 'network_provider' from source: set_fact 13731 1727203837.25104: Evaluated conditional (network_provider == "nm"): True 13731 1727203837.25374: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203837.25379: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203837.25486: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203837.27577: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203837.27624: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203837.27651: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203837.27678: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203837.27698: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203837.27757: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203837.27780: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203837.27799: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203837.27827: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203837.27839: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203837.27874: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203837.27892: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203837.27907: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203837.27934: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203837.27946: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203837.27977: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203837.27993: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203837.28009: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203837.28033: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203837.28043: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203837.28137: variable 'network_connections' from source: include params 13731 1727203837.28147: variable 'controller_profile' from source: play vars 13731 1727203837.28200: variable 'controller_profile' from source: play vars 13731 1727203837.28209: variable 'controller_device' from source: play vars 13731 1727203837.28253: variable 'controller_device' from source: play vars 13731 1727203837.28266: variable 'port1_profile' from source: play vars 13731 1727203837.28308: variable 'port1_profile' from source: play vars 13731 1727203837.28314: variable 'dhcp_interface1' from source: play vars 13731 1727203837.28357: variable 'dhcp_interface1' from source: play vars 13731 1727203837.28360: variable 'controller_profile' from source: play vars 13731 1727203837.28405: variable 'controller_profile' from source: play vars 13731 1727203837.28411: variable 'port2_profile' from source: play vars 13731 1727203837.28453: variable 'port2_profile' from source: play vars 13731 1727203837.28459: variable 'dhcp_interface2' from source: play vars 13731 1727203837.28505: variable 'dhcp_interface2' from source: play vars 13731 1727203837.28511: variable 'controller_profile' from source: play vars 13731 1727203837.28565: variable 'controller_profile' from source: play vars 13731 1727203837.28616: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203837.28726: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203837.28753: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203837.28777: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203837.28798: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203837.28831: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203837.28846: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203837.28899: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203837.28938: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203837.28972: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203837.29144: variable 'network_connections' from source: include params 13731 1727203837.29149: variable 'controller_profile' from source: play vars 13731 1727203837.29247: variable 'controller_profile' from source: play vars 13731 1727203837.29250: variable 'controller_device' from source: play vars 13731 1727203837.29310: variable 'controller_device' from source: play vars 13731 1727203837.29326: variable 'port1_profile' from source: play vars 13731 1727203837.29391: variable 'port1_profile' from source: play vars 13731 1727203837.29480: variable 'dhcp_interface1' from source: play vars 13731 1727203837.29484: variable 'dhcp_interface1' from source: play vars 13731 1727203837.29486: variable 'controller_profile' from source: play vars 13731 1727203837.29541: variable 'controller_profile' from source: play vars 13731 1727203837.29552: variable 'port2_profile' from source: play vars 13731 1727203837.29613: variable 'port2_profile' from source: play vars 13731 1727203837.29626: variable 'dhcp_interface2' from source: play vars 13731 1727203837.29690: variable 'dhcp_interface2' from source: play vars 13731 1727203837.29705: variable 'controller_profile' from source: play vars 13731 1727203837.29767: variable 'controller_profile' from source: play vars 13731 1727203837.29818: Evaluated conditional (__network_wpa_supplicant_required): False 13731 1727203837.29826: when evaluation is False, skipping this task 13731 1727203837.29833: _execute() done 13731 1727203837.29838: dumping result to json 13731 1727203837.29845: done dumping result, returning 13731 1727203837.29856: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant [028d2410-947f-82dc-c122-000000000284] 13731 1727203837.29869: sending task result for task 028d2410-947f-82dc-c122-000000000284 13731 1727203837.30094: done sending task result for task 028d2410-947f-82dc-c122-000000000284 13731 1727203837.30097: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wpa_supplicant_required", "skip_reason": "Conditional result was False" } 13731 1727203837.30172: no more pending results, returning what we have 13731 1727203837.30178: results queue empty 13731 1727203837.30179: checking for any_errors_fatal 13731 1727203837.30202: done checking for any_errors_fatal 13731 1727203837.30202: checking for max_fail_percentage 13731 1727203837.30204: done checking for max_fail_percentage 13731 1727203837.30205: checking to see if all hosts have failed and the running result is not ok 13731 1727203837.30205: done checking to see if all hosts have failed 13731 1727203837.30206: getting the remaining hosts for this loop 13731 1727203837.30208: done getting the remaining hosts for this loop 13731 1727203837.30211: getting the next task for host managed-node3 13731 1727203837.30217: done getting next task for host managed-node3 13731 1727203837.30221: ^ task is: TASK: fedora.linux_system_roles.network : Enable network service 13731 1727203837.30225: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203837.30239: getting variables 13731 1727203837.30240: in VariableManager get_vars() 13731 1727203837.30273: Calling all_inventory to load vars for managed-node3 13731 1727203837.30278: Calling groups_inventory to load vars for managed-node3 13731 1727203837.30280: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203837.30288: Calling all_plugins_play to load vars for managed-node3 13731 1727203837.30290: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203837.30292: Calling groups_plugins_play to load vars for managed-node3 13731 1727203837.31349: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203837.32512: done with get_vars() 13731 1727203837.32531: done getting variables 13731 1727203837.32581: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable network service] ************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:142 Tuesday 24 September 2024 14:50:37 -0400 (0:00:00.086) 0:00:15.557 ***** 13731 1727203837.32604: entering _queue_task() for managed-node3/service 13731 1727203837.32843: worker is 1 (out of 1 available) 13731 1727203837.32857: exiting _queue_task() for managed-node3/service 13731 1727203837.32871: done queuing things up, now waiting for results queue to drain 13731 1727203837.32873: waiting for pending results... 13731 1727203837.33044: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable network service 13731 1727203837.33148: in run() - task 028d2410-947f-82dc-c122-000000000285 13731 1727203837.33159: variable 'ansible_search_path' from source: unknown 13731 1727203837.33162: variable 'ansible_search_path' from source: unknown 13731 1727203837.33195: calling self._execute() 13731 1727203837.33261: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203837.33269: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203837.33279: variable 'omit' from source: magic vars 13731 1727203837.33552: variable 'ansible_distribution_major_version' from source: facts 13731 1727203837.33562: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203837.33642: variable 'network_provider' from source: set_fact 13731 1727203837.33652: Evaluated conditional (network_provider == "initscripts"): False 13731 1727203837.33656: when evaluation is False, skipping this task 13731 1727203837.33666: _execute() done 13731 1727203837.33669: dumping result to json 13731 1727203837.33672: done dumping result, returning 13731 1727203837.33675: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable network service [028d2410-947f-82dc-c122-000000000285] 13731 1727203837.33679: sending task result for task 028d2410-947f-82dc-c122-000000000285 13731 1727203837.33763: done sending task result for task 028d2410-947f-82dc-c122-000000000285 13731 1727203837.33767: WORKER PROCESS EXITING skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203837.33810: no more pending results, returning what we have 13731 1727203837.33814: results queue empty 13731 1727203837.33814: checking for any_errors_fatal 13731 1727203837.33825: done checking for any_errors_fatal 13731 1727203837.33826: checking for max_fail_percentage 13731 1727203837.33827: done checking for max_fail_percentage 13731 1727203837.33828: checking to see if all hosts have failed and the running result is not ok 13731 1727203837.33828: done checking to see if all hosts have failed 13731 1727203837.33829: getting the remaining hosts for this loop 13731 1727203837.33831: done getting the remaining hosts for this loop 13731 1727203837.33834: getting the next task for host managed-node3 13731 1727203837.33841: done getting next task for host managed-node3 13731 1727203837.33844: ^ task is: TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13731 1727203837.33855: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=19, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203837.33880: getting variables 13731 1727203837.33882: in VariableManager get_vars() 13731 1727203837.33921: Calling all_inventory to load vars for managed-node3 13731 1727203837.33924: Calling groups_inventory to load vars for managed-node3 13731 1727203837.33926: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203837.33938: Calling all_plugins_play to load vars for managed-node3 13731 1727203837.33940: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203837.33943: Calling groups_plugins_play to load vars for managed-node3 13731 1727203837.35239: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203837.37026: done with get_vars() 13731 1727203837.37050: done getting variables 13731 1727203837.37115: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ensure initscripts network file dependency is present] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:150 Tuesday 24 September 2024 14:50:37 -0400 (0:00:00.045) 0:00:15.602 ***** 13731 1727203837.37154: entering _queue_task() for managed-node3/copy 13731 1727203837.37504: worker is 1 (out of 1 available) 13731 1727203837.37517: exiting _queue_task() for managed-node3/copy 13731 1727203837.37529: done queuing things up, now waiting for results queue to drain 13731 1727203837.37531: waiting for pending results... 13731 1727203837.37907: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13731 1727203837.37979: in run() - task 028d2410-947f-82dc-c122-000000000286 13731 1727203837.38010: variable 'ansible_search_path' from source: unknown 13731 1727203837.38019: variable 'ansible_search_path' from source: unknown 13731 1727203837.38060: calling self._execute() 13731 1727203837.38160: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203837.38173: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203837.38218: variable 'omit' from source: magic vars 13731 1727203837.38601: variable 'ansible_distribution_major_version' from source: facts 13731 1727203837.38620: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203837.38766: variable 'network_provider' from source: set_fact 13731 1727203837.38874: Evaluated conditional (network_provider == "initscripts"): False 13731 1727203837.38879: when evaluation is False, skipping this task 13731 1727203837.38882: _execute() done 13731 1727203837.38885: dumping result to json 13731 1727203837.38887: done dumping result, returning 13731 1727203837.38891: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present [028d2410-947f-82dc-c122-000000000286] 13731 1727203837.38893: sending task result for task 028d2410-947f-82dc-c122-000000000286 13731 1727203837.38971: done sending task result for task 028d2410-947f-82dc-c122-000000000286 13731 1727203837.39081: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_provider == \"initscripts\"", "skip_reason": "Conditional result was False" } 13731 1727203837.39136: no more pending results, returning what we have 13731 1727203837.39141: results queue empty 13731 1727203837.39142: checking for any_errors_fatal 13731 1727203837.39149: done checking for any_errors_fatal 13731 1727203837.39150: checking for max_fail_percentage 13731 1727203837.39152: done checking for max_fail_percentage 13731 1727203837.39152: checking to see if all hosts have failed and the running result is not ok 13731 1727203837.39153: done checking to see if all hosts have failed 13731 1727203837.39154: getting the remaining hosts for this loop 13731 1727203837.39156: done getting the remaining hosts for this loop 13731 1727203837.39160: getting the next task for host managed-node3 13731 1727203837.39172: done getting next task for host managed-node3 13731 1727203837.39178: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13731 1727203837.39184: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=20, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203837.39200: getting variables 13731 1727203837.39202: in VariableManager get_vars() 13731 1727203837.39242: Calling all_inventory to load vars for managed-node3 13731 1727203837.39245: Calling groups_inventory to load vars for managed-node3 13731 1727203837.39248: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203837.39260: Calling all_plugins_play to load vars for managed-node3 13731 1727203837.39267: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203837.39270: Calling groups_plugins_play to load vars for managed-node3 13731 1727203837.40802: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203837.42453: done with get_vars() 13731 1727203837.42480: done getting variables TASK [fedora.linux_system_roles.network : Configure networking connection profiles] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 Tuesday 24 September 2024 14:50:37 -0400 (0:00:00.054) 0:00:15.656 ***** 13731 1727203837.42574: entering _queue_task() for managed-node3/fedora.linux_system_roles.network_connections 13731 1727203837.42577: Creating lock for fedora.linux_system_roles.network_connections 13731 1727203837.42907: worker is 1 (out of 1 available) 13731 1727203837.42921: exiting _queue_task() for managed-node3/fedora.linux_system_roles.network_connections 13731 1727203837.42934: done queuing things up, now waiting for results queue to drain 13731 1727203837.42936: waiting for pending results... 13731 1727203837.43304: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13731 1727203837.43357: in run() - task 028d2410-947f-82dc-c122-000000000287 13731 1727203837.43382: variable 'ansible_search_path' from source: unknown 13731 1727203837.43391: variable 'ansible_search_path' from source: unknown 13731 1727203837.43444: calling self._execute() 13731 1727203837.43559: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203837.43576: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203837.43592: variable 'omit' from source: magic vars 13731 1727203837.44018: variable 'ansible_distribution_major_version' from source: facts 13731 1727203837.44057: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203837.44063: variable 'omit' from source: magic vars 13731 1727203837.44122: variable 'omit' from source: magic vars 13731 1727203837.44385: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203837.46499: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203837.46574: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203837.46617: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203837.46655: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203837.46696: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203837.46783: variable 'network_provider' from source: set_fact 13731 1727203837.46920: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203837.46952: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203837.46986: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203837.47039: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203837.47064: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203837.47218: variable 'omit' from source: magic vars 13731 1727203837.47280: variable 'omit' from source: magic vars 13731 1727203837.47405: variable 'network_connections' from source: include params 13731 1727203837.47421: variable 'controller_profile' from source: play vars 13731 1727203837.47494: variable 'controller_profile' from source: play vars 13731 1727203837.47506: variable 'controller_device' from source: play vars 13731 1727203837.47577: variable 'controller_device' from source: play vars 13731 1727203837.47595: variable 'port1_profile' from source: play vars 13731 1727203837.47664: variable 'port1_profile' from source: play vars 13731 1727203837.47677: variable 'dhcp_interface1' from source: play vars 13731 1727203837.47738: variable 'dhcp_interface1' from source: play vars 13731 1727203837.47763: variable 'controller_profile' from source: play vars 13731 1727203837.47872: variable 'controller_profile' from source: play vars 13731 1727203837.47877: variable 'port2_profile' from source: play vars 13731 1727203837.47901: variable 'port2_profile' from source: play vars 13731 1727203837.47912: variable 'dhcp_interface2' from source: play vars 13731 1727203837.47980: variable 'dhcp_interface2' from source: play vars 13731 1727203837.47994: variable 'controller_profile' from source: play vars 13731 1727203837.48053: variable 'controller_profile' from source: play vars 13731 1727203837.48308: variable 'omit' from source: magic vars 13731 1727203837.48311: variable '__lsr_ansible_managed' from source: task vars 13731 1727203837.48369: variable '__lsr_ansible_managed' from source: task vars 13731 1727203837.48781: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 13731 1727203837.49097: Loaded config def from plugin (lookup/template) 13731 1727203837.49107: Loading LookupModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/template.py 13731 1727203837.49145: File lookup term: get_ansible_managed.j2 13731 1727203837.49181: variable 'ansible_search_path' from source: unknown 13731 1727203837.49185: evaluation_path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks 13731 1727203837.49188: search_path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 13731 1727203837.49202: variable 'ansible_search_path' from source: unknown 13731 1727203837.56785: variable 'ansible_managed' from source: unknown 13731 1727203837.57134: variable 'omit' from source: magic vars 13731 1727203837.57138: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203837.57351: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203837.57355: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203837.57357: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203837.57359: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203837.57365: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203837.57367: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203837.57369: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203837.57553: Set connection var ansible_pipelining to False 13731 1727203837.57782: Set connection var ansible_shell_type to sh 13731 1727203837.57785: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203837.57787: Set connection var ansible_connection to ssh 13731 1727203837.57799: Set connection var ansible_shell_executable to /bin/sh 13731 1727203837.57802: Set connection var ansible_timeout to 10 13731 1727203837.57804: variable 'ansible_shell_executable' from source: unknown 13731 1727203837.57806: variable 'ansible_connection' from source: unknown 13731 1727203837.57808: variable 'ansible_module_compression' from source: unknown 13731 1727203837.57810: variable 'ansible_shell_type' from source: unknown 13731 1727203837.57812: variable 'ansible_shell_executable' from source: unknown 13731 1727203837.57814: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203837.57816: variable 'ansible_pipelining' from source: unknown 13731 1727203837.57817: variable 'ansible_timeout' from source: unknown 13731 1727203837.57819: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203837.58108: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203837.58122: variable 'omit' from source: magic vars 13731 1727203837.58133: starting attempt loop 13731 1727203837.58138: running the handler 13731 1727203837.58153: _low_level_execute_command(): starting 13731 1727203837.58164: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203837.59598: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203837.59603: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203837.59799: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203837.59812: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203837.59868: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203837.61597: stdout chunk (state=3): >>>/root <<< 13731 1727203837.61797: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203837.62081: stderr chunk (state=3): >>><<< 13731 1727203837.62084: stdout chunk (state=3): >>><<< 13731 1727203837.62087: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203837.62089: _low_level_execute_command(): starting 13731 1727203837.62092: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354 `" && echo ansible-tmp-1727203837.6191628-14683-269180971665354="` echo /root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354 `" ) && sleep 0' 13731 1727203837.63410: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203837.63525: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203837.63682: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203837.63894: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203837.63958: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203837.65844: stdout chunk (state=3): >>>ansible-tmp-1727203837.6191628-14683-269180971665354=/root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354 <<< 13731 1727203837.66062: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203837.66066: stdout chunk (state=3): >>><<< 13731 1727203837.66074: stderr chunk (state=3): >>><<< 13731 1727203837.66096: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203837.6191628-14683-269180971665354=/root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203837.66146: variable 'ansible_module_compression' from source: unknown 13731 1727203837.66199: ANSIBALLZ: Using lock for fedora.linux_system_roles.network_connections 13731 1727203837.66202: ANSIBALLZ: Acquiring lock 13731 1727203837.66205: ANSIBALLZ: Lock acquired: 140078451140624 13731 1727203837.66207: ANSIBALLZ: Creating module 13731 1727203837.97343: ANSIBALLZ: Writing module into payload 13731 1727203837.97652: ANSIBALLZ: Writing module 13731 1727203837.97676: ANSIBALLZ: Renaming module 13731 1727203837.97683: ANSIBALLZ: Done creating module 13731 1727203837.97708: variable 'ansible_facts' from source: unknown 13731 1727203837.97821: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/AnsiballZ_network_connections.py 13731 1727203837.98049: Sending initial data 13731 1727203837.98053: Sent initial data (168 bytes) 13731 1727203837.98652: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203837.98722: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203837.98741: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203837.98764: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203837.98833: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203838.00585: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203838.00589: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203838.00591: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpduq9zqfl /root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/AnsiballZ_network_connections.py <<< 13731 1727203838.00594: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/AnsiballZ_network_connections.py" <<< 13731 1727203838.00597: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpduq9zqfl" to remote "/root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/AnsiballZ_network_connections.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/AnsiballZ_network_connections.py" <<< 13731 1727203838.01685: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203838.01746: stderr chunk (state=3): >>><<< 13731 1727203838.01756: stdout chunk (state=3): >>><<< 13731 1727203838.01784: done transferring module to remote 13731 1727203838.01790: _low_level_execute_command(): starting 13731 1727203838.01796: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/ /root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/AnsiballZ_network_connections.py && sleep 0' 13731 1727203838.02404: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203838.02417: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203838.02428: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203838.02441: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203838.02454: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203838.02460: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203838.02470: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.02487: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203838.02495: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203838.02501: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203838.02525: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.02636: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203838.02641: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203838.02677: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203838.04450: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203838.04454: stdout chunk (state=3): >>><<< 13731 1727203838.04457: stderr chunk (state=3): >>><<< 13731 1727203838.04487: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203838.04490: _low_level_execute_command(): starting 13731 1727203838.04587: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/AnsiballZ_network_connections.py && sleep 0' 13731 1727203838.05306: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203838.05325: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203838.05348: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203838.05537: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.05760: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203838.05763: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203838.05787: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203838.05858: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203838.45006: stdout chunk (state=3): >>> {"changed": true, "warnings": [], "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d (not-active)\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "802.3ad", "ad_actor_sys_prio": 65535, "ad_actor_system": "00:00:5e:00:53:5d", "ad_select": "stable", "ad_user_port_key": 1023, "all_ports_active": true, "downdelay": 0, "lacp_rate": "slow", "lp_interval": 128, "miimon": 110, "min_links": 0, "num_grat_arp": 64, "primary_reselect": "better", "resend_igmp": 225, "updelay": 0, "use_carrier": true, "xmit_hash_policy": "encap2+3"}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "802.3ad", "ad_actor_sys_prio": 65535, "ad_actor_system": "00:00:5e:00:53:5d", "ad_select": "stable", "ad_user_port_key": 1023, "all_ports_active": true, "downdelay": 0, "lacp_rate": "slow", "lp_interval": 128, "miimon": 110, "min_links": 0, "num_grat_arp": 64, "primary_reselect": "better", "resend_igmp": 225, "updelay": 0, "use_carrier": true, "xmit_hash_policy": "encap2+3"}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} <<< 13731 1727203838.47023: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203838.47053: stderr chunk (state=3): >>><<< 13731 1727203838.47057: stdout chunk (state=3): >>><<< 13731 1727203838.47078: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "warnings": [], "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d (not-active)\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "802.3ad", "ad_actor_sys_prio": 65535, "ad_actor_system": "00:00:5e:00:53:5d", "ad_select": "stable", "ad_user_port_key": 1023, "all_ports_active": true, "downdelay": 0, "lacp_rate": "slow", "lp_interval": 128, "miimon": 110, "min_links": 0, "num_grat_arp": 64, "primary_reselect": "better", "resend_igmp": 225, "updelay": 0, "use_carrier": true, "xmit_hash_policy": "encap2+3"}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "802.3ad", "ad_actor_sys_prio": 65535, "ad_actor_system": "00:00:5e:00:53:5d", "ad_select": "stable", "ad_user_port_key": 1023, "all_ports_active": true, "downdelay": 0, "lacp_rate": "slow", "lp_interval": 128, "miimon": 110, "min_links": 0, "num_grat_arp": 64, "primary_reselect": "better", "resend_igmp": 225, "updelay": 0, "use_carrier": true, "xmit_hash_policy": "encap2+3"}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203838.47134: done with _execute_module (fedora.linux_system_roles.network_connections, {'provider': 'nm', 'connections': [{'name': 'bond0', 'state': 'up', 'type': 'bond', 'interface_name': 'nm-bond', 'bond': {'mode': '802.3ad', 'ad_actor_sys_prio': 65535, 'ad_actor_system': '00:00:5e:00:53:5d', 'ad_select': 'stable', 'ad_user_port_key': 1023, 'all_ports_active': True, 'downdelay': 0, 'lacp_rate': 'slow', 'lp_interval': 128, 'miimon': 110, 'min_links': 0, 'num_grat_arp': 64, 'primary_reselect': 'better', 'resend_igmp': 225, 'updelay': 0, 'use_carrier': True, 'xmit_hash_policy': 'encap2+3'}, 'ip': {'route_metric4': 65535}}, {'name': 'bond0.0', 'state': 'up', 'type': 'ethernet', 'interface_name': 'test1', 'controller': 'bond0'}, {'name': 'bond0.1', 'state': 'up', 'type': 'ethernet', 'interface_name': 'test2', 'controller': 'bond0'}], '__header': '#\n# Ansible managed\n#\n# system_role:network\n', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'fedora.linux_system_roles.network_connections', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203838.47143: _low_level_execute_command(): starting 13731 1727203838.47147: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203837.6191628-14683-269180971665354/ > /dev/null 2>&1 && sleep 0' 13731 1727203838.47602: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203838.47606: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203838.47608: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.47610: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203838.47612: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203838.47614: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.47665: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203838.47668: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203838.47671: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203838.47710: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203838.49585: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203838.49606: stderr chunk (state=3): >>><<< 13731 1727203838.49609: stdout chunk (state=3): >>><<< 13731 1727203838.49621: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203838.49627: handler run complete 13731 1727203838.49660: attempt loop complete, returning result 13731 1727203838.49663: _execute() done 13731 1727203838.49667: dumping result to json 13731 1727203838.49676: done dumping result, returning 13731 1727203838.49688: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking connection profiles [028d2410-947f-82dc-c122-000000000287] 13731 1727203838.49691: sending task result for task 028d2410-947f-82dc-c122-000000000287 13731 1727203838.49811: done sending task result for task 028d2410-947f-82dc-c122-000000000287 13731 1727203838.49813: WORKER PROCESS EXITING changed: [managed-node3] => { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "bond": { "ad_actor_sys_prio": 65535, "ad_actor_system": "00:00:5e:00:53:5d", "ad_select": "stable", "ad_user_port_key": 1023, "all_ports_active": true, "downdelay": 0, "lacp_rate": "slow", "lp_interval": 128, "miimon": 110, "min_links": 0, "mode": "802.3ad", "num_grat_arp": 64, "primary_reselect": "better", "resend_igmp": 225, "updelay": 0, "use_carrier": true, "xmit_hash_policy": "encap2+3" }, "interface_name": "nm-bond", "ip": { "route_metric4": 65535 }, "name": "bond0", "state": "up", "type": "bond" }, { "controller": "bond0", "interface_name": "test1", "name": "bond0.0", "state": "up", "type": "ethernet" }, { "controller": "bond0", "interface_name": "test2", "name": "bond0.1", "state": "up", "type": "ethernet" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true } STDERR: [007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 [008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d [009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d [010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 (is-modified) [011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d (not-active) [012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d (not-active) 13731 1727203838.49960: no more pending results, returning what we have 13731 1727203838.49964: results queue empty 13731 1727203838.49965: checking for any_errors_fatal 13731 1727203838.49974: done checking for any_errors_fatal 13731 1727203838.49975: checking for max_fail_percentage 13731 1727203838.49983: done checking for max_fail_percentage 13731 1727203838.49983: checking to see if all hosts have failed and the running result is not ok 13731 1727203838.49984: done checking to see if all hosts have failed 13731 1727203838.49985: getting the remaining hosts for this loop 13731 1727203838.49987: done getting the remaining hosts for this loop 13731 1727203838.49990: getting the next task for host managed-node3 13731 1727203838.49996: done getting next task for host managed-node3 13731 1727203838.49999: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking state 13731 1727203838.50003: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=21, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203838.50013: getting variables 13731 1727203838.50014: in VariableManager get_vars() 13731 1727203838.50047: Calling all_inventory to load vars for managed-node3 13731 1727203838.50050: Calling groups_inventory to load vars for managed-node3 13731 1727203838.50052: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203838.50061: Calling all_plugins_play to load vars for managed-node3 13731 1727203838.50063: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203838.50065: Calling groups_plugins_play to load vars for managed-node3 13731 1727203838.51009: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203838.51860: done with get_vars() 13731 1727203838.51878: done getting variables TASK [fedora.linux_system_roles.network : Configure networking state] ********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:171 Tuesday 24 September 2024 14:50:38 -0400 (0:00:01.093) 0:00:16.750 ***** 13731 1727203838.51940: entering _queue_task() for managed-node3/fedora.linux_system_roles.network_state 13731 1727203838.51942: Creating lock for fedora.linux_system_roles.network_state 13731 1727203838.52171: worker is 1 (out of 1 available) 13731 1727203838.52187: exiting _queue_task() for managed-node3/fedora.linux_system_roles.network_state 13731 1727203838.52199: done queuing things up, now waiting for results queue to drain 13731 1727203838.52201: waiting for pending results... 13731 1727203838.52372: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking state 13731 1727203838.52463: in run() - task 028d2410-947f-82dc-c122-000000000288 13731 1727203838.52478: variable 'ansible_search_path' from source: unknown 13731 1727203838.52482: variable 'ansible_search_path' from source: unknown 13731 1727203838.52510: calling self._execute() 13731 1727203838.52580: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.52584: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.52592: variable 'omit' from source: magic vars 13731 1727203838.52870: variable 'ansible_distribution_major_version' from source: facts 13731 1727203838.52882: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203838.52962: variable 'network_state' from source: role '' defaults 13731 1727203838.52978: Evaluated conditional (network_state != {}): False 13731 1727203838.52982: when evaluation is False, skipping this task 13731 1727203838.52985: _execute() done 13731 1727203838.52987: dumping result to json 13731 1727203838.52990: done dumping result, returning 13731 1727203838.52992: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking state [028d2410-947f-82dc-c122-000000000288] 13731 1727203838.52998: sending task result for task 028d2410-947f-82dc-c122-000000000288 13731 1727203838.53080: done sending task result for task 028d2410-947f-82dc-c122-000000000288 13731 1727203838.53083: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203838.53131: no more pending results, returning what we have 13731 1727203838.53135: results queue empty 13731 1727203838.53136: checking for any_errors_fatal 13731 1727203838.53152: done checking for any_errors_fatal 13731 1727203838.53153: checking for max_fail_percentage 13731 1727203838.53155: done checking for max_fail_percentage 13731 1727203838.53156: checking to see if all hosts have failed and the running result is not ok 13731 1727203838.53157: done checking to see if all hosts have failed 13731 1727203838.53157: getting the remaining hosts for this loop 13731 1727203838.53159: done getting the remaining hosts for this loop 13731 1727203838.53162: getting the next task for host managed-node3 13731 1727203838.53168: done getting next task for host managed-node3 13731 1727203838.53172: ^ task is: TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13731 1727203838.53178: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=22, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203838.53192: getting variables 13731 1727203838.53194: in VariableManager get_vars() 13731 1727203838.53223: Calling all_inventory to load vars for managed-node3 13731 1727203838.53226: Calling groups_inventory to load vars for managed-node3 13731 1727203838.53228: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203838.53235: Calling all_plugins_play to load vars for managed-node3 13731 1727203838.53237: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203838.53240: Calling groups_plugins_play to load vars for managed-node3 13731 1727203838.54000: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203838.54881: done with get_vars() 13731 1727203838.54896: done getting variables 13731 1727203838.54939: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show stderr messages for the network_connections] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:177 Tuesday 24 September 2024 14:50:38 -0400 (0:00:00.030) 0:00:16.780 ***** 13731 1727203838.54964: entering _queue_task() for managed-node3/debug 13731 1727203838.55179: worker is 1 (out of 1 available) 13731 1727203838.55193: exiting _queue_task() for managed-node3/debug 13731 1727203838.55204: done queuing things up, now waiting for results queue to drain 13731 1727203838.55206: waiting for pending results... 13731 1727203838.55403: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13731 1727203838.55483: in run() - task 028d2410-947f-82dc-c122-000000000289 13731 1727203838.55495: variable 'ansible_search_path' from source: unknown 13731 1727203838.55499: variable 'ansible_search_path' from source: unknown 13731 1727203838.55528: calling self._execute() 13731 1727203838.55598: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.55602: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.55610: variable 'omit' from source: magic vars 13731 1727203838.56108: variable 'ansible_distribution_major_version' from source: facts 13731 1727203838.56111: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203838.56114: variable 'omit' from source: magic vars 13731 1727203838.56116: variable 'omit' from source: magic vars 13731 1727203838.56118: variable 'omit' from source: magic vars 13731 1727203838.56120: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203838.56151: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203838.56182: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203838.56205: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203838.56223: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203838.56258: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203838.56270: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.56280: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.56379: Set connection var ansible_pipelining to False 13731 1727203838.56390: Set connection var ansible_shell_type to sh 13731 1727203838.56399: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203838.56407: Set connection var ansible_connection to ssh 13731 1727203838.56415: Set connection var ansible_shell_executable to /bin/sh 13731 1727203838.56425: Set connection var ansible_timeout to 10 13731 1727203838.56447: variable 'ansible_shell_executable' from source: unknown 13731 1727203838.56455: variable 'ansible_connection' from source: unknown 13731 1727203838.56468: variable 'ansible_module_compression' from source: unknown 13731 1727203838.56474: variable 'ansible_shell_type' from source: unknown 13731 1727203838.56482: variable 'ansible_shell_executable' from source: unknown 13731 1727203838.56489: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.56495: variable 'ansible_pipelining' from source: unknown 13731 1727203838.56501: variable 'ansible_timeout' from source: unknown 13731 1727203838.56507: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.56665: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203838.56697: variable 'omit' from source: magic vars 13731 1727203838.56706: starting attempt loop 13731 1727203838.56713: running the handler 13731 1727203838.56859: variable '__network_connections_result' from source: set_fact 13731 1727203838.56965: handler run complete 13731 1727203838.56994: attempt loop complete, returning result 13731 1727203838.57005: _execute() done 13731 1727203838.57012: dumping result to json 13731 1727203838.57081: done dumping result, returning 13731 1727203838.57084: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections [028d2410-947f-82dc-c122-000000000289] 13731 1727203838.57086: sending task result for task 028d2410-947f-82dc-c122-000000000289 13731 1727203838.57154: done sending task result for task 028d2410-947f-82dc-c122-000000000289 13731 1727203838.57159: WORKER PROCESS EXITING ok: [managed-node3] => { "__network_connections_result.stderr_lines": [ "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8", "[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d", "[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d", "[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 (is-modified)", "[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d (not-active)", "[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d (not-active)" ] } 13731 1727203838.57224: no more pending results, returning what we have 13731 1727203838.57228: results queue empty 13731 1727203838.57229: checking for any_errors_fatal 13731 1727203838.57236: done checking for any_errors_fatal 13731 1727203838.57237: checking for max_fail_percentage 13731 1727203838.57238: done checking for max_fail_percentage 13731 1727203838.57239: checking to see if all hosts have failed and the running result is not ok 13731 1727203838.57239: done checking to see if all hosts have failed 13731 1727203838.57240: getting the remaining hosts for this loop 13731 1727203838.57242: done getting the remaining hosts for this loop 13731 1727203838.57245: getting the next task for host managed-node3 13731 1727203838.57252: done getting next task for host managed-node3 13731 1727203838.57255: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13731 1727203838.57260: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=23, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203838.57272: getting variables 13731 1727203838.57274: in VariableManager get_vars() 13731 1727203838.57307: Calling all_inventory to load vars for managed-node3 13731 1727203838.57310: Calling groups_inventory to load vars for managed-node3 13731 1727203838.57312: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203838.57321: Calling all_plugins_play to load vars for managed-node3 13731 1727203838.57330: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203838.57332: Calling groups_plugins_play to load vars for managed-node3 13731 1727203838.58618: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203838.59999: done with get_vars() 13731 1727203838.60021: done getting variables 13731 1727203838.60077: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_connections] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:181 Tuesday 24 September 2024 14:50:38 -0400 (0:00:00.051) 0:00:16.832 ***** 13731 1727203838.60110: entering _queue_task() for managed-node3/debug 13731 1727203838.60405: worker is 1 (out of 1 available) 13731 1727203838.60416: exiting _queue_task() for managed-node3/debug 13731 1727203838.60428: done queuing things up, now waiting for results queue to drain 13731 1727203838.60429: waiting for pending results... 13731 1727203838.60894: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13731 1727203838.60899: in run() - task 028d2410-947f-82dc-c122-00000000028a 13731 1727203838.60902: variable 'ansible_search_path' from source: unknown 13731 1727203838.60905: variable 'ansible_search_path' from source: unknown 13731 1727203838.60908: calling self._execute() 13731 1727203838.61001: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.61012: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.61029: variable 'omit' from source: magic vars 13731 1727203838.61381: variable 'ansible_distribution_major_version' from source: facts 13731 1727203838.61458: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203838.61461: variable 'omit' from source: magic vars 13731 1727203838.61477: variable 'omit' from source: magic vars 13731 1727203838.61513: variable 'omit' from source: magic vars 13731 1727203838.61555: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203838.61599: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203838.61620: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203838.61640: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203838.61656: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203838.61696: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203838.61703: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.61710: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.61890: Set connection var ansible_pipelining to False 13731 1727203838.61893: Set connection var ansible_shell_type to sh 13731 1727203838.61895: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203838.61896: Set connection var ansible_connection to ssh 13731 1727203838.61898: Set connection var ansible_shell_executable to /bin/sh 13731 1727203838.61900: Set connection var ansible_timeout to 10 13731 1727203838.61901: variable 'ansible_shell_executable' from source: unknown 13731 1727203838.61903: variable 'ansible_connection' from source: unknown 13731 1727203838.61905: variable 'ansible_module_compression' from source: unknown 13731 1727203838.61907: variable 'ansible_shell_type' from source: unknown 13731 1727203838.61909: variable 'ansible_shell_executable' from source: unknown 13731 1727203838.61910: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.61912: variable 'ansible_pipelining' from source: unknown 13731 1727203838.61913: variable 'ansible_timeout' from source: unknown 13731 1727203838.61915: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.62025: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203838.62044: variable 'omit' from source: magic vars 13731 1727203838.62053: starting attempt loop 13731 1727203838.62059: running the handler 13731 1727203838.62115: variable '__network_connections_result' from source: set_fact 13731 1727203838.62197: variable '__network_connections_result' from source: set_fact 13731 1727203838.62415: handler run complete 13731 1727203838.62461: attempt loop complete, returning result 13731 1727203838.62468: _execute() done 13731 1727203838.62474: dumping result to json 13731 1727203838.62487: done dumping result, returning 13731 1727203838.62497: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections [028d2410-947f-82dc-c122-00000000028a] 13731 1727203838.62541: sending task result for task 028d2410-947f-82dc-c122-00000000028a ok: [managed-node3] => { "__network_connections_result": { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "bond": { "ad_actor_sys_prio": 65535, "ad_actor_system": "00:00:5e:00:53:5d", "ad_select": "stable", "ad_user_port_key": 1023, "all_ports_active": true, "downdelay": 0, "lacp_rate": "slow", "lp_interval": 128, "miimon": 110, "min_links": 0, "mode": "802.3ad", "num_grat_arp": 64, "primary_reselect": "better", "resend_igmp": 225, "updelay": 0, "use_carrier": true, "xmit_hash_policy": "encap2+3" }, "interface_name": "nm-bond", "ip": { "route_metric4": 65535 }, "name": "bond0", "state": "up", "type": "bond" }, { "controller": "bond0", "interface_name": "test1", "name": "bond0.0", "state": "up", "type": "ethernet" }, { "controller": "bond0", "interface_name": "test2", "name": "bond0.1", "state": "up", "type": "ethernet" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true, "failed": false, "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d (not-active)\n", "stderr_lines": [ "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8", "[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d", "[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d", "[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 (is-modified)", "[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d (not-active)", "[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d (not-active)" ] } } 13731 1727203838.62762: no more pending results, returning what we have 13731 1727203838.62767: results queue empty 13731 1727203838.62768: checking for any_errors_fatal 13731 1727203838.62776: done checking for any_errors_fatal 13731 1727203838.62777: checking for max_fail_percentage 13731 1727203838.62779: done checking for max_fail_percentage 13731 1727203838.62779: checking to see if all hosts have failed and the running result is not ok 13731 1727203838.62780: done checking to see if all hosts have failed 13731 1727203838.62781: getting the remaining hosts for this loop 13731 1727203838.62783: done getting the remaining hosts for this loop 13731 1727203838.62786: getting the next task for host managed-node3 13731 1727203838.62797: done getting next task for host managed-node3 13731 1727203838.62800: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13731 1727203838.62805: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=24, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203838.62816: getting variables 13731 1727203838.62817: in VariableManager get_vars() 13731 1727203838.62852: Calling all_inventory to load vars for managed-node3 13731 1727203838.62855: Calling groups_inventory to load vars for managed-node3 13731 1727203838.62857: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203838.62867: Calling all_plugins_play to load vars for managed-node3 13731 1727203838.62870: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203838.62873: Calling groups_plugins_play to load vars for managed-node3 13731 1727203838.63130: done sending task result for task 028d2410-947f-82dc-c122-00000000028a 13731 1727203838.63134: WORKER PROCESS EXITING 13731 1727203838.64400: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203838.65956: done with get_vars() 13731 1727203838.65983: done getting variables 13731 1727203838.66040: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_state] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:186 Tuesday 24 September 2024 14:50:38 -0400 (0:00:00.059) 0:00:16.891 ***** 13731 1727203838.66071: entering _queue_task() for managed-node3/debug 13731 1727203838.66381: worker is 1 (out of 1 available) 13731 1727203838.66393: exiting _queue_task() for managed-node3/debug 13731 1727203838.66404: done queuing things up, now waiting for results queue to drain 13731 1727203838.66406: waiting for pending results... 13731 1727203838.66677: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13731 1727203838.67182: in run() - task 028d2410-947f-82dc-c122-00000000028b 13731 1727203838.67186: variable 'ansible_search_path' from source: unknown 13731 1727203838.67189: variable 'ansible_search_path' from source: unknown 13731 1727203838.67191: calling self._execute() 13731 1727203838.67194: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.67196: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.67199: variable 'omit' from source: magic vars 13731 1727203838.67937: variable 'ansible_distribution_major_version' from source: facts 13731 1727203838.67956: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203838.68307: variable 'network_state' from source: role '' defaults 13731 1727203838.68323: Evaluated conditional (network_state != {}): False 13731 1727203838.68332: when evaluation is False, skipping this task 13731 1727203838.68339: _execute() done 13731 1727203838.68346: dumping result to json 13731 1727203838.68352: done dumping result, returning 13731 1727203838.68362: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state [028d2410-947f-82dc-c122-00000000028b] 13731 1727203838.68372: sending task result for task 028d2410-947f-82dc-c122-00000000028b 13731 1727203838.68488: done sending task result for task 028d2410-947f-82dc-c122-00000000028b skipping: [managed-node3] => { "false_condition": "network_state != {}" } 13731 1727203838.68555: no more pending results, returning what we have 13731 1727203838.68560: results queue empty 13731 1727203838.68561: checking for any_errors_fatal 13731 1727203838.68579: done checking for any_errors_fatal 13731 1727203838.68580: checking for max_fail_percentage 13731 1727203838.68582: done checking for max_fail_percentage 13731 1727203838.68583: checking to see if all hosts have failed and the running result is not ok 13731 1727203838.68584: done checking to see if all hosts have failed 13731 1727203838.68584: getting the remaining hosts for this loop 13731 1727203838.68587: done getting the remaining hosts for this loop 13731 1727203838.68590: getting the next task for host managed-node3 13731 1727203838.68599: done getting next task for host managed-node3 13731 1727203838.68603: ^ task is: TASK: fedora.linux_system_roles.network : Re-test connectivity 13731 1727203838.68608: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=25, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203838.68624: getting variables 13731 1727203838.68626: in VariableManager get_vars() 13731 1727203838.68665: Calling all_inventory to load vars for managed-node3 13731 1727203838.68668: Calling groups_inventory to load vars for managed-node3 13731 1727203838.68671: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203838.68987: Calling all_plugins_play to load vars for managed-node3 13731 1727203838.68991: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203838.68996: Calling groups_plugins_play to load vars for managed-node3 13731 1727203838.70204: WORKER PROCESS EXITING 13731 1727203838.71683: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203838.73352: done with get_vars() 13731 1727203838.73379: done getting variables TASK [fedora.linux_system_roles.network : Re-test connectivity] **************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:192 Tuesday 24 September 2024 14:50:38 -0400 (0:00:00.074) 0:00:16.965 ***** 13731 1727203838.73480: entering _queue_task() for managed-node3/ping 13731 1727203838.73482: Creating lock for ping 13731 1727203838.73816: worker is 1 (out of 1 available) 13731 1727203838.73829: exiting _queue_task() for managed-node3/ping 13731 1727203838.73841: done queuing things up, now waiting for results queue to drain 13731 1727203838.73843: waiting for pending results... 13731 1727203838.74116: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Re-test connectivity 13731 1727203838.74342: in run() - task 028d2410-947f-82dc-c122-00000000028c 13731 1727203838.74371: variable 'ansible_search_path' from source: unknown 13731 1727203838.74383: variable 'ansible_search_path' from source: unknown 13731 1727203838.74456: calling self._execute() 13731 1727203838.74629: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.74641: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.74657: variable 'omit' from source: magic vars 13731 1727203838.75424: variable 'ansible_distribution_major_version' from source: facts 13731 1727203838.75501: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203838.75513: variable 'omit' from source: magic vars 13731 1727203838.75696: variable 'omit' from source: magic vars 13731 1727203838.75813: variable 'omit' from source: magic vars 13731 1727203838.75940: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203838.76056: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203838.76086: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203838.76159: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203838.76315: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203838.76319: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203838.76322: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.76324: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.76558: Set connection var ansible_pipelining to False 13731 1727203838.76600: Set connection var ansible_shell_type to sh 13731 1727203838.76611: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203838.76619: Set connection var ansible_connection to ssh 13731 1727203838.76628: Set connection var ansible_shell_executable to /bin/sh 13731 1727203838.76643: Set connection var ansible_timeout to 10 13731 1727203838.76671: variable 'ansible_shell_executable' from source: unknown 13731 1727203838.76685: variable 'ansible_connection' from source: unknown 13731 1727203838.76693: variable 'ansible_module_compression' from source: unknown 13731 1727203838.76700: variable 'ansible_shell_type' from source: unknown 13731 1727203838.76705: variable 'ansible_shell_executable' from source: unknown 13731 1727203838.76750: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203838.76753: variable 'ansible_pipelining' from source: unknown 13731 1727203838.76755: variable 'ansible_timeout' from source: unknown 13731 1727203838.76757: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203838.76956: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203838.76984: variable 'omit' from source: magic vars 13731 1727203838.77077: starting attempt loop 13731 1727203838.77083: running the handler 13731 1727203838.77087: _low_level_execute_command(): starting 13731 1727203838.77090: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203838.77622: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203838.77633: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.77656: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203838.77660: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.77718: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203838.77724: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203838.77727: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203838.77767: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203838.79922: stdout chunk (state=3): >>>/root <<< 13731 1727203838.79926: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203838.79928: stdout chunk (state=3): >>><<< 13731 1727203838.79930: stderr chunk (state=3): >>><<< 13731 1727203838.79934: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203838.79936: _low_level_execute_command(): starting 13731 1727203838.79939: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250 `" && echo ansible-tmp-1727203838.7979193-14743-252299572548250="` echo /root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250 `" ) && sleep 0' 13731 1727203838.80335: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203838.80339: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203838.80395: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.80427: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203838.80434: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203838.80466: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203838.82348: stdout chunk (state=3): >>>ansible-tmp-1727203838.7979193-14743-252299572548250=/root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250 <<< 13731 1727203838.82710: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203838.82714: stdout chunk (state=3): >>><<< 13731 1727203838.82717: stderr chunk (state=3): >>><<< 13731 1727203838.82720: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203838.7979193-14743-252299572548250=/root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203838.82985: variable 'ansible_module_compression' from source: unknown 13731 1727203838.82988: ANSIBALLZ: Using lock for ping 13731 1727203838.82990: ANSIBALLZ: Acquiring lock 13731 1727203838.82992: ANSIBALLZ: Lock acquired: 140078455234896 13731 1727203838.82994: ANSIBALLZ: Creating module 13731 1727203838.95490: ANSIBALLZ: Writing module into payload 13731 1727203838.95528: ANSIBALLZ: Writing module 13731 1727203838.95544: ANSIBALLZ: Renaming module 13731 1727203838.95550: ANSIBALLZ: Done creating module 13731 1727203838.95569: variable 'ansible_facts' from source: unknown 13731 1727203838.95613: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/AnsiballZ_ping.py 13731 1727203838.95710: Sending initial data 13731 1727203838.95714: Sent initial data (153 bytes) 13731 1727203838.96157: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203838.96160: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203838.96165: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.96167: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203838.96169: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203838.96171: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.96223: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203838.96228: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203838.96230: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203838.96272: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203838.97885: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" <<< 13731 1727203838.97892: stderr chunk (state=3): >>>debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203838.97916: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203838.97950: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp1kwk85r0 /root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/AnsiballZ_ping.py <<< 13731 1727203838.97965: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/AnsiballZ_ping.py" <<< 13731 1727203838.97985: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp1kwk85r0" to remote "/root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/AnsiballZ_ping.py" <<< 13731 1727203838.97990: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/AnsiballZ_ping.py" <<< 13731 1727203838.98481: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203838.98523: stderr chunk (state=3): >>><<< 13731 1727203838.98526: stdout chunk (state=3): >>><<< 13731 1727203838.98543: done transferring module to remote 13731 1727203838.98552: _low_level_execute_command(): starting 13731 1727203838.98556: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/ /root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/AnsiballZ_ping.py && sleep 0' 13731 1727203838.98967: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203838.98971: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203838.99002: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203838.99005: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203838.99008: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203838.99013: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203838.99058: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203838.99064: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203838.99082: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203838.99111: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203839.00829: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203839.00852: stderr chunk (state=3): >>><<< 13731 1727203839.00855: stdout chunk (state=3): >>><<< 13731 1727203839.00874: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203839.00880: _low_level_execute_command(): starting 13731 1727203839.00883: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/AnsiballZ_ping.py && sleep 0' 13731 1727203839.01325: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203839.01329: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203839.01331: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203839.01333: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203839.01381: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203839.01385: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203839.01390: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203839.01433: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203839.16143: stdout chunk (state=3): >>> {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} <<< 13731 1727203839.17411: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203839.17414: stdout chunk (state=3): >>><<< 13731 1727203839.17417: stderr chunk (state=3): >>><<< 13731 1727203839.17432: _low_level_execute_command() done: rc=0, stdout= {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203839.17553: done with _execute_module (ping, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ping', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203839.17557: _low_level_execute_command(): starting 13731 1727203839.17559: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203838.7979193-14743-252299572548250/ > /dev/null 2>&1 && sleep 0' 13731 1727203839.18296: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203839.18318: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203839.18337: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203839.18410: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203839.20204: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203839.20262: stderr chunk (state=3): >>><<< 13731 1727203839.20301: stdout chunk (state=3): >>><<< 13731 1727203839.20327: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203839.20361: handler run complete 13731 1727203839.20581: attempt loop complete, returning result 13731 1727203839.20584: _execute() done 13731 1727203839.20586: dumping result to json 13731 1727203839.20589: done dumping result, returning 13731 1727203839.20591: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Re-test connectivity [028d2410-947f-82dc-c122-00000000028c] 13731 1727203839.20593: sending task result for task 028d2410-947f-82dc-c122-00000000028c 13731 1727203839.20661: done sending task result for task 028d2410-947f-82dc-c122-00000000028c 13731 1727203839.20665: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "ping": "pong" } 13731 1727203839.20731: no more pending results, returning what we have 13731 1727203839.20735: results queue empty 13731 1727203839.20736: checking for any_errors_fatal 13731 1727203839.20744: done checking for any_errors_fatal 13731 1727203839.20744: checking for max_fail_percentage 13731 1727203839.20746: done checking for max_fail_percentage 13731 1727203839.20747: checking to see if all hosts have failed and the running result is not ok 13731 1727203839.20748: done checking to see if all hosts have failed 13731 1727203839.20748: getting the remaining hosts for this loop 13731 1727203839.20750: done getting the remaining hosts for this loop 13731 1727203839.20754: getting the next task for host managed-node3 13731 1727203839.20764: done getting next task for host managed-node3 13731 1727203839.20767: ^ task is: TASK: meta (role_complete) 13731 1727203839.20772: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203839.20787: getting variables 13731 1727203839.20789: in VariableManager get_vars() 13731 1727203839.20829: Calling all_inventory to load vars for managed-node3 13731 1727203839.20832: Calling groups_inventory to load vars for managed-node3 13731 1727203839.20834: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203839.20845: Calling all_plugins_play to load vars for managed-node3 13731 1727203839.20848: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203839.20852: Calling groups_plugins_play to load vars for managed-node3 13731 1727203839.23312: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203839.25046: done with get_vars() 13731 1727203839.25129: done getting variables 13731 1727203839.25311: done queuing things up, now waiting for results queue to drain 13731 1727203839.25313: results queue empty 13731 1727203839.25314: checking for any_errors_fatal 13731 1727203839.25317: done checking for any_errors_fatal 13731 1727203839.25318: checking for max_fail_percentage 13731 1727203839.25319: done checking for max_fail_percentage 13731 1727203839.25320: checking to see if all hosts have failed and the running result is not ok 13731 1727203839.25321: done checking to see if all hosts have failed 13731 1727203839.25321: getting the remaining hosts for this loop 13731 1727203839.25322: done getting the remaining hosts for this loop 13731 1727203839.25325: getting the next task for host managed-node3 13731 1727203839.25330: done getting next task for host managed-node3 13731 1727203839.25337: ^ task is: TASK: Show result 13731 1727203839.25340: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203839.25343: getting variables 13731 1727203839.25344: in VariableManager get_vars() 13731 1727203839.25356: Calling all_inventory to load vars for managed-node3 13731 1727203839.25358: Calling groups_inventory to load vars for managed-node3 13731 1727203839.25360: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203839.25366: Calling all_plugins_play to load vars for managed-node3 13731 1727203839.25368: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203839.25371: Calling groups_plugins_play to load vars for managed-node3 13731 1727203839.27513: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203839.29527: done with get_vars() 13731 1727203839.29552: done getting variables 13731 1727203839.29598: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Show result] ************************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile.yml:46 Tuesday 24 September 2024 14:50:39 -0400 (0:00:00.561) 0:00:17.527 ***** 13731 1727203839.29629: entering _queue_task() for managed-node3/debug 13731 1727203839.29957: worker is 1 (out of 1 available) 13731 1727203839.29970: exiting _queue_task() for managed-node3/debug 13731 1727203839.29984: done queuing things up, now waiting for results queue to drain 13731 1727203839.29986: waiting for pending results... 13731 1727203839.30267: running TaskExecutor() for managed-node3/TASK: Show result 13731 1727203839.30363: in run() - task 028d2410-947f-82dc-c122-0000000001c6 13731 1727203839.30368: variable 'ansible_search_path' from source: unknown 13731 1727203839.30370: variable 'ansible_search_path' from source: unknown 13731 1727203839.30394: calling self._execute() 13731 1727203839.30678: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.30884: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.30888: variable 'omit' from source: magic vars 13731 1727203839.31482: variable 'ansible_distribution_major_version' from source: facts 13731 1727203839.31514: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203839.31525: variable 'omit' from source: magic vars 13731 1727203839.31572: variable 'omit' from source: magic vars 13731 1727203839.31763: variable 'omit' from source: magic vars 13731 1727203839.31809: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203839.31866: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203839.31955: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203839.31980: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203839.32037: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203839.32098: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203839.32105: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.32114: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.32241: Set connection var ansible_pipelining to False 13731 1727203839.32251: Set connection var ansible_shell_type to sh 13731 1727203839.32260: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203839.32278: Set connection var ansible_connection to ssh 13731 1727203839.32290: Set connection var ansible_shell_executable to /bin/sh 13731 1727203839.32301: Set connection var ansible_timeout to 10 13731 1727203839.32328: variable 'ansible_shell_executable' from source: unknown 13731 1727203839.32336: variable 'ansible_connection' from source: unknown 13731 1727203839.32344: variable 'ansible_module_compression' from source: unknown 13731 1727203839.32350: variable 'ansible_shell_type' from source: unknown 13731 1727203839.32358: variable 'ansible_shell_executable' from source: unknown 13731 1727203839.32367: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.32381: variable 'ansible_pipelining' from source: unknown 13731 1727203839.32400: variable 'ansible_timeout' from source: unknown 13731 1727203839.32408: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.32552: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203839.32591: variable 'omit' from source: magic vars 13731 1727203839.32594: starting attempt loop 13731 1727203839.32596: running the handler 13731 1727203839.32681: variable '__network_connections_result' from source: set_fact 13731 1727203839.32757: variable '__network_connections_result' from source: set_fact 13731 1727203839.33014: handler run complete 13731 1727203839.33062: attempt loop complete, returning result 13731 1727203839.33132: _execute() done 13731 1727203839.33136: dumping result to json 13731 1727203839.33139: done dumping result, returning 13731 1727203839.33141: done running TaskExecutor() for managed-node3/TASK: Show result [028d2410-947f-82dc-c122-0000000001c6] 13731 1727203839.33143: sending task result for task 028d2410-947f-82dc-c122-0000000001c6 ok: [managed-node3] => { "__network_connections_result": { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "bond": { "ad_actor_sys_prio": 65535, "ad_actor_system": "00:00:5e:00:53:5d", "ad_select": "stable", "ad_user_port_key": 1023, "all_ports_active": true, "downdelay": 0, "lacp_rate": "slow", "lp_interval": 128, "miimon": 110, "min_links": 0, "mode": "802.3ad", "num_grat_arp": 64, "primary_reselect": "better", "resend_igmp": 225, "updelay": 0, "use_carrier": true, "xmit_hash_policy": "encap2+3" }, "interface_name": "nm-bond", "ip": { "route_metric4": 65535 }, "name": "bond0", "state": "up", "type": "bond" }, { "controller": "bond0", "interface_name": "test1", "name": "bond0.0", "state": "up", "type": "ethernet" }, { "controller": "bond0", "interface_name": "test2", "name": "bond0.1", "state": "up", "type": "ethernet" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true, "failed": false, "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d (not-active)\n", "stderr_lines": [ "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8", "[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d", "[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d", "[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 49770e6d-8624-49fd-b707-6e339710b2f8 (is-modified)", "[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 4dc9735e-3039-485e-9eca-4fa95cb0bf9d (not-active)", "[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 52cf19c0-ba8d-4c54-a280-415ff537447d (not-active)" ] } } 13731 1727203839.33333: no more pending results, returning what we have 13731 1727203839.33339: results queue empty 13731 1727203839.33340: checking for any_errors_fatal 13731 1727203839.33341: done checking for any_errors_fatal 13731 1727203839.33342: checking for max_fail_percentage 13731 1727203839.33344: done checking for max_fail_percentage 13731 1727203839.33344: checking to see if all hosts have failed and the running result is not ok 13731 1727203839.33345: done checking to see if all hosts have failed 13731 1727203839.33346: getting the remaining hosts for this loop 13731 1727203839.33349: done getting the remaining hosts for this loop 13731 1727203839.33352: getting the next task for host managed-node3 13731 1727203839.33363: done getting next task for host managed-node3 13731 1727203839.33366: ^ task is: TASK: Asserts 13731 1727203839.33370: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203839.33374: getting variables 13731 1727203839.33377: in VariableManager get_vars() 13731 1727203839.33407: Calling all_inventory to load vars for managed-node3 13731 1727203839.33410: Calling groups_inventory to load vars for managed-node3 13731 1727203839.33413: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203839.33424: Calling all_plugins_play to load vars for managed-node3 13731 1727203839.33427: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203839.33430: Calling groups_plugins_play to load vars for managed-node3 13731 1727203839.33951: done sending task result for task 028d2410-947f-82dc-c122-0000000001c6 13731 1727203839.33954: WORKER PROCESS EXITING 13731 1727203839.35111: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203839.36765: done with get_vars() 13731 1727203839.36791: done getting variables TASK [Asserts] ***************************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:36 Tuesday 24 September 2024 14:50:39 -0400 (0:00:00.072) 0:00:17.599 ***** 13731 1727203839.36889: entering _queue_task() for managed-node3/include_tasks 13731 1727203839.37238: worker is 1 (out of 1 available) 13731 1727203839.37257: exiting _queue_task() for managed-node3/include_tasks 13731 1727203839.37270: done queuing things up, now waiting for results queue to drain 13731 1727203839.37272: waiting for pending results... 13731 1727203839.37557: running TaskExecutor() for managed-node3/TASK: Asserts 13731 1727203839.37671: in run() - task 028d2410-947f-82dc-c122-00000000008d 13731 1727203839.37695: variable 'ansible_search_path' from source: unknown 13731 1727203839.37707: variable 'ansible_search_path' from source: unknown 13731 1727203839.37753: variable 'lsr_assert' from source: include params 13731 1727203839.37965: variable 'lsr_assert' from source: include params 13731 1727203839.38040: variable 'omit' from source: magic vars 13731 1727203839.38245: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.38247: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.38250: variable 'omit' from source: magic vars 13731 1727203839.38432: variable 'ansible_distribution_major_version' from source: facts 13731 1727203839.38447: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203839.38463: variable 'item' from source: unknown 13731 1727203839.38529: variable 'item' from source: unknown 13731 1727203839.38570: variable 'item' from source: unknown 13731 1727203839.38634: variable 'item' from source: unknown 13731 1727203839.38983: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.38987: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.38989: variable 'omit' from source: magic vars 13731 1727203839.39022: variable 'ansible_distribution_major_version' from source: facts 13731 1727203839.39033: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203839.39042: variable 'item' from source: unknown 13731 1727203839.39110: variable 'item' from source: unknown 13731 1727203839.39144: variable 'item' from source: unknown 13731 1727203839.39215: variable 'item' from source: unknown 13731 1727203839.39418: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.39421: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.39423: variable 'omit' from source: magic vars 13731 1727203839.39551: variable 'ansible_distribution_major_version' from source: facts 13731 1727203839.39567: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203839.39638: variable 'item' from source: unknown 13731 1727203839.39652: variable 'item' from source: unknown 13731 1727203839.39692: variable 'item' from source: unknown 13731 1727203839.39765: variable 'item' from source: unknown 13731 1727203839.39880: dumping result to json 13731 1727203839.39884: done dumping result, returning 13731 1727203839.39886: done running TaskExecutor() for managed-node3/TASK: Asserts [028d2410-947f-82dc-c122-00000000008d] 13731 1727203839.39890: sending task result for task 028d2410-947f-82dc-c122-00000000008d 13731 1727203839.40006: done sending task result for task 028d2410-947f-82dc-c122-00000000008d 13731 1727203839.40009: WORKER PROCESS EXITING 13731 1727203839.40101: no more pending results, returning what we have 13731 1727203839.40107: in VariableManager get_vars() 13731 1727203839.40146: Calling all_inventory to load vars for managed-node3 13731 1727203839.40149: Calling groups_inventory to load vars for managed-node3 13731 1727203839.40153: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203839.40170: Calling all_plugins_play to load vars for managed-node3 13731 1727203839.40173: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203839.40178: Calling groups_plugins_play to load vars for managed-node3 13731 1727203839.41683: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203839.43159: done with get_vars() 13731 1727203839.43186: variable 'ansible_search_path' from source: unknown 13731 1727203839.43188: variable 'ansible_search_path' from source: unknown 13731 1727203839.43232: variable 'ansible_search_path' from source: unknown 13731 1727203839.43233: variable 'ansible_search_path' from source: unknown 13731 1727203839.43267: variable 'ansible_search_path' from source: unknown 13731 1727203839.43268: variable 'ansible_search_path' from source: unknown 13731 1727203839.43299: we have included files to process 13731 1727203839.43301: generating all_blocks data 13731 1727203839.43303: done generating all_blocks data 13731 1727203839.43307: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_controller_device_present.yml 13731 1727203839.43308: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_controller_device_present.yml 13731 1727203839.43311: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_controller_device_present.yml 13731 1727203839.43465: in VariableManager get_vars() 13731 1727203839.43487: done with get_vars() 13731 1727203839.43493: variable 'item' from source: include params 13731 1727203839.43596: variable 'item' from source: include params 13731 1727203839.43622: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml statically imported: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml 13731 1727203839.43693: in VariableManager get_vars() 13731 1727203839.43713: done with get_vars() 13731 1727203839.43843: done processing included file 13731 1727203839.43845: iterating over new_blocks loaded from include file 13731 1727203839.43846: in VariableManager get_vars() 13731 1727203839.43861: done with get_vars() 13731 1727203839.43865: filtering new block on tags 13731 1727203839.43906: done filtering new block on tags 13731 1727203839.43909: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_controller_device_present.yml for managed-node3 => (item=tasks/assert_controller_device_present.yml) 13731 1727203839.43914: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_port_profile_present.yml 13731 1727203839.43915: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_port_profile_present.yml 13731 1727203839.43917: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_port_profile_present.yml 13731 1727203839.44038: in VariableManager get_vars() 13731 1727203839.44054: done with get_vars() 13731 1727203839.44068: done processing included file 13731 1727203839.44070: iterating over new_blocks loaded from include file 13731 1727203839.44071: in VariableManager get_vars() 13731 1727203839.44084: done with get_vars() 13731 1727203839.44085: filtering new block on tags 13731 1727203839.44105: done filtering new block on tags 13731 1727203839.44107: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_port_profile_present.yml for managed-node3 => (item=tasks/assert_bond_port_profile_present.yml) 13731 1727203839.44110: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml 13731 1727203839.44111: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml 13731 1727203839.44118: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml 13731 1727203839.44411: in VariableManager get_vars() 13731 1727203839.44427: done with get_vars() 13731 1727203839.44466: in VariableManager get_vars() 13731 1727203839.44483: done with get_vars() 13731 1727203839.44495: done processing included file 13731 1727203839.44496: iterating over new_blocks loaded from include file 13731 1727203839.44497: in VariableManager get_vars() 13731 1727203839.44508: done with get_vars() 13731 1727203839.44510: filtering new block on tags 13731 1727203839.44545: done filtering new block on tags 13731 1727203839.44547: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml for managed-node3 => (item=tasks/assert_bond_options.yml) 13731 1727203839.44550: extending task lists for all hosts with included blocks 13731 1727203839.45989: done extending task lists 13731 1727203839.45990: done processing included files 13731 1727203839.45991: results queue empty 13731 1727203839.45992: checking for any_errors_fatal 13731 1727203839.45998: done checking for any_errors_fatal 13731 1727203839.45998: checking for max_fail_percentage 13731 1727203839.45999: done checking for max_fail_percentage 13731 1727203839.46000: checking to see if all hosts have failed and the running result is not ok 13731 1727203839.46001: done checking to see if all hosts have failed 13731 1727203839.46002: getting the remaining hosts for this loop 13731 1727203839.46003: done getting the remaining hosts for this loop 13731 1727203839.46005: getting the next task for host managed-node3 13731 1727203839.46010: done getting next task for host managed-node3 13731 1727203839.46012: ^ task is: TASK: Include the task 'get_interface_stat.yml' 13731 1727203839.46015: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203839.46017: getting variables 13731 1727203839.46018: in VariableManager get_vars() 13731 1727203839.46026: Calling all_inventory to load vars for managed-node3 13731 1727203839.46028: Calling groups_inventory to load vars for managed-node3 13731 1727203839.46030: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203839.46035: Calling all_plugins_play to load vars for managed-node3 13731 1727203839.46038: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203839.46040: Calling groups_plugins_play to load vars for managed-node3 13731 1727203839.47158: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203839.48671: done with get_vars() 13731 1727203839.48701: done getting variables TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:3 Tuesday 24 September 2024 14:50:39 -0400 (0:00:00.118) 0:00:17.718 ***** 13731 1727203839.48791: entering _queue_task() for managed-node3/include_tasks 13731 1727203839.49157: worker is 1 (out of 1 available) 13731 1727203839.49170: exiting _queue_task() for managed-node3/include_tasks 13731 1727203839.49186: done queuing things up, now waiting for results queue to drain 13731 1727203839.49187: waiting for pending results... 13731 1727203839.49597: running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' 13731 1727203839.49608: in run() - task 028d2410-947f-82dc-c122-0000000003f5 13731 1727203839.49627: variable 'ansible_search_path' from source: unknown 13731 1727203839.49635: variable 'ansible_search_path' from source: unknown 13731 1727203839.49679: calling self._execute() 13731 1727203839.49782: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.49795: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.49811: variable 'omit' from source: magic vars 13731 1727203839.50199: variable 'ansible_distribution_major_version' from source: facts 13731 1727203839.50216: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203839.50226: _execute() done 13731 1727203839.50237: dumping result to json 13731 1727203839.50346: done dumping result, returning 13731 1727203839.50349: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' [028d2410-947f-82dc-c122-0000000003f5] 13731 1727203839.50352: sending task result for task 028d2410-947f-82dc-c122-0000000003f5 13731 1727203839.50427: done sending task result for task 028d2410-947f-82dc-c122-0000000003f5 13731 1727203839.50430: WORKER PROCESS EXITING 13731 1727203839.50483: no more pending results, returning what we have 13731 1727203839.50489: in VariableManager get_vars() 13731 1727203839.50532: Calling all_inventory to load vars for managed-node3 13731 1727203839.50536: Calling groups_inventory to load vars for managed-node3 13731 1727203839.50539: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203839.50553: Calling all_plugins_play to load vars for managed-node3 13731 1727203839.50556: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203839.50559: Calling groups_plugins_play to load vars for managed-node3 13731 1727203839.52499: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203839.59425: done with get_vars() 13731 1727203839.59446: variable 'ansible_search_path' from source: unknown 13731 1727203839.59447: variable 'ansible_search_path' from source: unknown 13731 1727203839.59488: we have included files to process 13731 1727203839.59490: generating all_blocks data 13731 1727203839.59491: done generating all_blocks data 13731 1727203839.59492: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203839.59493: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203839.59495: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203839.59655: done processing included file 13731 1727203839.59657: iterating over new_blocks loaded from include file 13731 1727203839.59659: in VariableManager get_vars() 13731 1727203839.59679: done with get_vars() 13731 1727203839.59680: filtering new block on tags 13731 1727203839.59708: done filtering new block on tags 13731 1727203839.59710: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node3 13731 1727203839.59714: extending task lists for all hosts with included blocks 13731 1727203839.59911: done extending task lists 13731 1727203839.59912: done processing included files 13731 1727203839.59913: results queue empty 13731 1727203839.59914: checking for any_errors_fatal 13731 1727203839.59916: done checking for any_errors_fatal 13731 1727203839.59917: checking for max_fail_percentage 13731 1727203839.59918: done checking for max_fail_percentage 13731 1727203839.59919: checking to see if all hosts have failed and the running result is not ok 13731 1727203839.59919: done checking to see if all hosts have failed 13731 1727203839.59920: getting the remaining hosts for this loop 13731 1727203839.59921: done getting the remaining hosts for this loop 13731 1727203839.59923: getting the next task for host managed-node3 13731 1727203839.59928: done getting next task for host managed-node3 13731 1727203839.59930: ^ task is: TASK: Get stat for interface {{ interface }} 13731 1727203839.59933: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203839.59935: getting variables 13731 1727203839.59936: in VariableManager get_vars() 13731 1727203839.59944: Calling all_inventory to load vars for managed-node3 13731 1727203839.59946: Calling groups_inventory to load vars for managed-node3 13731 1727203839.59949: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203839.59954: Calling all_plugins_play to load vars for managed-node3 13731 1727203839.59956: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203839.59959: Calling groups_plugins_play to load vars for managed-node3 13731 1727203839.61104: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203839.62673: done with get_vars() 13731 1727203839.62695: done getting variables 13731 1727203839.62836: variable 'interface' from source: task vars 13731 1727203839.62840: variable 'controller_device' from source: play vars 13731 1727203839.62901: variable 'controller_device' from source: play vars TASK [Get stat for interface nm-bond] ****************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 Tuesday 24 September 2024 14:50:39 -0400 (0:00:00.141) 0:00:17.860 ***** 13731 1727203839.62930: entering _queue_task() for managed-node3/stat 13731 1727203839.63297: worker is 1 (out of 1 available) 13731 1727203839.63308: exiting _queue_task() for managed-node3/stat 13731 1727203839.63321: done queuing things up, now waiting for results queue to drain 13731 1727203839.63323: waiting for pending results... 13731 1727203839.63608: running TaskExecutor() for managed-node3/TASK: Get stat for interface nm-bond 13731 1727203839.63782: in run() - task 028d2410-947f-82dc-c122-0000000004af 13731 1727203839.63785: variable 'ansible_search_path' from source: unknown 13731 1727203839.63787: variable 'ansible_search_path' from source: unknown 13731 1727203839.63812: calling self._execute() 13731 1727203839.64183: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.64188: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.64191: variable 'omit' from source: magic vars 13731 1727203839.64957: variable 'ansible_distribution_major_version' from source: facts 13731 1727203839.64996: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203839.65010: variable 'omit' from source: magic vars 13731 1727203839.65305: variable 'omit' from source: magic vars 13731 1727203839.65413: variable 'interface' from source: task vars 13731 1727203839.65452: variable 'controller_device' from source: play vars 13731 1727203839.65572: variable 'controller_device' from source: play vars 13731 1727203839.65734: variable 'omit' from source: magic vars 13731 1727203839.65830: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203839.65952: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203839.65983: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203839.66025: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203839.66282: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203839.66285: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203839.66287: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.66289: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.66439: Set connection var ansible_pipelining to False 13731 1727203839.66451: Set connection var ansible_shell_type to sh 13731 1727203839.66488: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203839.66501: Set connection var ansible_connection to ssh 13731 1727203839.66715: Set connection var ansible_shell_executable to /bin/sh 13731 1727203839.66719: Set connection var ansible_timeout to 10 13731 1727203839.66721: variable 'ansible_shell_executable' from source: unknown 13731 1727203839.66723: variable 'ansible_connection' from source: unknown 13731 1727203839.66725: variable 'ansible_module_compression' from source: unknown 13731 1727203839.66727: variable 'ansible_shell_type' from source: unknown 13731 1727203839.66729: variable 'ansible_shell_executable' from source: unknown 13731 1727203839.66731: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203839.66734: variable 'ansible_pipelining' from source: unknown 13731 1727203839.66736: variable 'ansible_timeout' from source: unknown 13731 1727203839.66738: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203839.67483: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203839.67487: variable 'omit' from source: magic vars 13731 1727203839.67490: starting attempt loop 13731 1727203839.67493: running the handler 13731 1727203839.67495: _low_level_execute_command(): starting 13731 1727203839.67497: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203839.68712: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203839.68730: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203839.68744: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203839.68788: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203839.68896: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203839.68916: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203839.68975: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203839.70649: stdout chunk (state=3): >>>/root <<< 13731 1727203839.70788: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203839.70793: stdout chunk (state=3): >>><<< 13731 1727203839.70802: stderr chunk (state=3): >>><<< 13731 1727203839.70826: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203839.70842: _low_level_execute_command(): starting 13731 1727203839.70845: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679 `" && echo ansible-tmp-1727203839.7082505-14789-41911517812679="` echo /root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679 `" ) && sleep 0' 13731 1727203839.72091: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203839.72200: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203839.72213: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203839.72279: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203839.74142: stdout chunk (state=3): >>>ansible-tmp-1727203839.7082505-14789-41911517812679=/root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679 <<< 13731 1727203839.74293: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203839.74296: stdout chunk (state=3): >>><<< 13731 1727203839.74298: stderr chunk (state=3): >>><<< 13731 1727203839.74548: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203839.7082505-14789-41911517812679=/root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203839.74608: variable 'ansible_module_compression' from source: unknown 13731 1727203839.74673: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13731 1727203839.74723: variable 'ansible_facts' from source: unknown 13731 1727203839.74828: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/AnsiballZ_stat.py 13731 1727203839.75045: Sending initial data 13731 1727203839.75048: Sent initial data (152 bytes) 13731 1727203839.75624: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203839.75639: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203839.75652: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203839.75688: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203839.75788: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203839.75810: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203839.75901: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203839.77397: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203839.77499: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203839.77542: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp1dao9hcn /root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/AnsiballZ_stat.py <<< 13731 1727203839.77545: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/AnsiballZ_stat.py" <<< 13731 1727203839.77572: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp1dao9hcn" to remote "/root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/AnsiballZ_stat.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/AnsiballZ_stat.py" <<< 13731 1727203839.78720: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203839.78835: stderr chunk (state=3): >>><<< 13731 1727203839.78838: stdout chunk (state=3): >>><<< 13731 1727203839.78840: done transferring module to remote 13731 1727203839.78842: _low_level_execute_command(): starting 13731 1727203839.78845: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/ /root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/AnsiballZ_stat.py && sleep 0' 13731 1727203839.80020: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203839.80293: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203839.80319: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203839.80328: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203839.80358: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203839.82073: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203839.82118: stderr chunk (state=3): >>><<< 13731 1727203839.82124: stdout chunk (state=3): >>><<< 13731 1727203839.82144: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203839.82152: _low_level_execute_command(): starting 13731 1727203839.82155: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/AnsiballZ_stat.py && sleep 0' 13731 1727203839.82782: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203839.82785: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203839.82788: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203839.82791: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203839.82793: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203839.82796: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203839.82798: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203839.82800: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203839.82802: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203839.82805: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203839.82807: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203839.82809: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203839.82811: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203839.82825: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203839.82828: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203839.82835: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203839.82906: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203839.82935: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203839.82939: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203839.82996: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203839.97987: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/nm-bond", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 28313, "dev": 23, "nlink": 1, "atime": 1727203838.3119485, "mtime": 1727203838.3119485, "ctime": 1727203838.3119485, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/nm-bond", "lnk_target": "../../devices/virtual/net/nm-bond", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/nm-bond", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13731 1727203839.99141: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203839.99161: stderr chunk (state=3): >>><<< 13731 1727203839.99164: stdout chunk (state=3): >>><<< 13731 1727203839.99186: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/nm-bond", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 28313, "dev": 23, "nlink": 1, "atime": 1727203838.3119485, "mtime": 1727203838.3119485, "ctime": 1727203838.3119485, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/nm-bond", "lnk_target": "../../devices/virtual/net/nm-bond", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/nm-bond", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203839.99226: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/sys/class/net/nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203839.99235: _low_level_execute_command(): starting 13731 1727203839.99238: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203839.7082505-14789-41911517812679/ > /dev/null 2>&1 && sleep 0' 13731 1727203839.99641: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203839.99682: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203839.99685: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203839.99687: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203839.99690: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203839.99697: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203839.99735: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203839.99742: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203839.99744: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203839.99774: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.01537: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.01562: stderr chunk (state=3): >>><<< 13731 1727203840.01568: stdout chunk (state=3): >>><<< 13731 1727203840.01583: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203840.01589: handler run complete 13731 1727203840.01623: attempt loop complete, returning result 13731 1727203840.01626: _execute() done 13731 1727203840.01629: dumping result to json 13731 1727203840.01633: done dumping result, returning 13731 1727203840.01641: done running TaskExecutor() for managed-node3/TASK: Get stat for interface nm-bond [028d2410-947f-82dc-c122-0000000004af] 13731 1727203840.01645: sending task result for task 028d2410-947f-82dc-c122-0000000004af 13731 1727203840.01752: done sending task result for task 028d2410-947f-82dc-c122-0000000004af 13731 1727203840.01754: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "stat": { "atime": 1727203838.3119485, "block_size": 4096, "blocks": 0, "ctime": 1727203838.3119485, "dev": 23, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 28313, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": true, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/sys/devices/virtual/net/nm-bond", "lnk_target": "../../devices/virtual/net/nm-bond", "mode": "0777", "mtime": 1727203838.3119485, "nlink": 1, "path": "/sys/class/net/nm-bond", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "wgrp": true, "woth": true, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 13731 1727203840.01838: no more pending results, returning what we have 13731 1727203840.01842: results queue empty 13731 1727203840.01843: checking for any_errors_fatal 13731 1727203840.01845: done checking for any_errors_fatal 13731 1727203840.01845: checking for max_fail_percentage 13731 1727203840.01847: done checking for max_fail_percentage 13731 1727203840.01848: checking to see if all hosts have failed and the running result is not ok 13731 1727203840.01849: done checking to see if all hosts have failed 13731 1727203840.01849: getting the remaining hosts for this loop 13731 1727203840.01851: done getting the remaining hosts for this loop 13731 1727203840.01855: getting the next task for host managed-node3 13731 1727203840.01864: done getting next task for host managed-node3 13731 1727203840.01867: ^ task is: TASK: Assert that the interface is present - '{{ interface }}' 13731 1727203840.01871: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203840.01882: getting variables 13731 1727203840.01884: in VariableManager get_vars() 13731 1727203840.01915: Calling all_inventory to load vars for managed-node3 13731 1727203840.01917: Calling groups_inventory to load vars for managed-node3 13731 1727203840.01920: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.01930: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.01932: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.01934: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.02789: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.03640: done with get_vars() 13731 1727203840.03654: done getting variables 13731 1727203840.03699: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203840.03787: variable 'interface' from source: task vars 13731 1727203840.03790: variable 'controller_device' from source: play vars 13731 1727203840.03832: variable 'controller_device' from source: play vars TASK [Assert that the interface is present - 'nm-bond'] ************************ task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:5 Tuesday 24 September 2024 14:50:40 -0400 (0:00:00.409) 0:00:18.269 ***** 13731 1727203840.03856: entering _queue_task() for managed-node3/assert 13731 1727203840.04079: worker is 1 (out of 1 available) 13731 1727203840.04093: exiting _queue_task() for managed-node3/assert 13731 1727203840.04106: done queuing things up, now waiting for results queue to drain 13731 1727203840.04107: waiting for pending results... 13731 1727203840.04297: running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'nm-bond' 13731 1727203840.04390: in run() - task 028d2410-947f-82dc-c122-0000000003f6 13731 1727203840.04400: variable 'ansible_search_path' from source: unknown 13731 1727203840.04403: variable 'ansible_search_path' from source: unknown 13731 1727203840.04437: calling self._execute() 13731 1727203840.04512: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.04517: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.04526: variable 'omit' from source: magic vars 13731 1727203840.04784: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.04795: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.04800: variable 'omit' from source: magic vars 13731 1727203840.04838: variable 'omit' from source: magic vars 13731 1727203840.04907: variable 'interface' from source: task vars 13731 1727203840.04910: variable 'controller_device' from source: play vars 13731 1727203840.04955: variable 'controller_device' from source: play vars 13731 1727203840.04971: variable 'omit' from source: magic vars 13731 1727203840.05003: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203840.05029: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203840.05045: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203840.05059: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203840.05074: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203840.05096: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203840.05099: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.05102: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.05166: Set connection var ansible_pipelining to False 13731 1727203840.05184: Set connection var ansible_shell_type to sh 13731 1727203840.05187: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203840.05190: Set connection var ansible_connection to ssh 13731 1727203840.05192: Set connection var ansible_shell_executable to /bin/sh 13731 1727203840.05199: Set connection var ansible_timeout to 10 13731 1727203840.05214: variable 'ansible_shell_executable' from source: unknown 13731 1727203840.05217: variable 'ansible_connection' from source: unknown 13731 1727203840.05220: variable 'ansible_module_compression' from source: unknown 13731 1727203840.05222: variable 'ansible_shell_type' from source: unknown 13731 1727203840.05224: variable 'ansible_shell_executable' from source: unknown 13731 1727203840.05227: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.05229: variable 'ansible_pipelining' from source: unknown 13731 1727203840.05232: variable 'ansible_timeout' from source: unknown 13731 1727203840.05234: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.05335: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203840.05343: variable 'omit' from source: magic vars 13731 1727203840.05348: starting attempt loop 13731 1727203840.05351: running the handler 13731 1727203840.05439: variable 'interface_stat' from source: set_fact 13731 1727203840.05454: Evaluated conditional (interface_stat.stat.exists): True 13731 1727203840.05460: handler run complete 13731 1727203840.05472: attempt loop complete, returning result 13731 1727203840.05477: _execute() done 13731 1727203840.05479: dumping result to json 13731 1727203840.05482: done dumping result, returning 13731 1727203840.05487: done running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'nm-bond' [028d2410-947f-82dc-c122-0000000003f6] 13731 1727203840.05493: sending task result for task 028d2410-947f-82dc-c122-0000000003f6 13731 1727203840.05572: done sending task result for task 028d2410-947f-82dc-c122-0000000003f6 13731 1727203840.05577: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203840.05621: no more pending results, returning what we have 13731 1727203840.05625: results queue empty 13731 1727203840.05626: checking for any_errors_fatal 13731 1727203840.05637: done checking for any_errors_fatal 13731 1727203840.05637: checking for max_fail_percentage 13731 1727203840.05639: done checking for max_fail_percentage 13731 1727203840.05640: checking to see if all hosts have failed and the running result is not ok 13731 1727203840.05640: done checking to see if all hosts have failed 13731 1727203840.05641: getting the remaining hosts for this loop 13731 1727203840.05643: done getting the remaining hosts for this loop 13731 1727203840.05646: getting the next task for host managed-node3 13731 1727203840.05656: done getting next task for host managed-node3 13731 1727203840.05658: ^ task is: TASK: Include the task 'assert_profile_present.yml' 13731 1727203840.05663: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203840.05667: getting variables 13731 1727203840.05668: in VariableManager get_vars() 13731 1727203840.05699: Calling all_inventory to load vars for managed-node3 13731 1727203840.05701: Calling groups_inventory to load vars for managed-node3 13731 1727203840.05705: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.05713: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.05715: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.05718: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.06472: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.07335: done with get_vars() 13731 1727203840.07349: done getting variables TASK [Include the task 'assert_profile_present.yml'] *************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_port_profile_present.yml:3 Tuesday 24 September 2024 14:50:40 -0400 (0:00:00.035) 0:00:18.305 ***** 13731 1727203840.07418: entering _queue_task() for managed-node3/include_tasks 13731 1727203840.07634: worker is 1 (out of 1 available) 13731 1727203840.07648: exiting _queue_task() for managed-node3/include_tasks 13731 1727203840.07661: done queuing things up, now waiting for results queue to drain 13731 1727203840.07662: waiting for pending results... 13731 1727203840.07836: running TaskExecutor() for managed-node3/TASK: Include the task 'assert_profile_present.yml' 13731 1727203840.07907: in run() - task 028d2410-947f-82dc-c122-0000000003fb 13731 1727203840.07920: variable 'ansible_search_path' from source: unknown 13731 1727203840.07924: variable 'ansible_search_path' from source: unknown 13731 1727203840.07957: variable 'controller_profile' from source: play vars 13731 1727203840.08095: variable 'controller_profile' from source: play vars 13731 1727203840.08112: variable 'port1_profile' from source: play vars 13731 1727203840.08155: variable 'port1_profile' from source: play vars 13731 1727203840.08161: variable 'port2_profile' from source: play vars 13731 1727203840.08209: variable 'port2_profile' from source: play vars 13731 1727203840.08222: variable 'omit' from source: magic vars 13731 1727203840.08329: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.08335: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.08343: variable 'omit' from source: magic vars 13731 1727203840.08511: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.08519: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.08543: variable 'bond_port_profile' from source: unknown 13731 1727203840.08591: variable 'bond_port_profile' from source: unknown 13731 1727203840.08712: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.08715: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.08718: variable 'omit' from source: magic vars 13731 1727203840.08795: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.08798: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.08819: variable 'bond_port_profile' from source: unknown 13731 1727203840.08863: variable 'bond_port_profile' from source: unknown 13731 1727203840.08927: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.08931: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.08943: variable 'omit' from source: magic vars 13731 1727203840.09036: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.09041: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.09067: variable 'bond_port_profile' from source: unknown 13731 1727203840.09109: variable 'bond_port_profile' from source: unknown 13731 1727203840.09176: dumping result to json 13731 1727203840.09180: done dumping result, returning 13731 1727203840.09183: done running TaskExecutor() for managed-node3/TASK: Include the task 'assert_profile_present.yml' [028d2410-947f-82dc-c122-0000000003fb] 13731 1727203840.09184: sending task result for task 028d2410-947f-82dc-c122-0000000003fb 13731 1727203840.09218: done sending task result for task 028d2410-947f-82dc-c122-0000000003fb 13731 1727203840.09221: WORKER PROCESS EXITING 13731 1727203840.09246: no more pending results, returning what we have 13731 1727203840.09251: in VariableManager get_vars() 13731 1727203840.09288: Calling all_inventory to load vars for managed-node3 13731 1727203840.09291: Calling groups_inventory to load vars for managed-node3 13731 1727203840.09295: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.09307: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.09309: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.09312: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.10178: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.11018: done with get_vars() 13731 1727203840.11030: variable 'ansible_search_path' from source: unknown 13731 1727203840.11031: variable 'ansible_search_path' from source: unknown 13731 1727203840.11037: variable 'item' from source: include params 13731 1727203840.11110: variable 'item' from source: include params 13731 1727203840.11136: variable 'ansible_search_path' from source: unknown 13731 1727203840.11137: variable 'ansible_search_path' from source: unknown 13731 1727203840.11141: variable 'item' from source: include params 13731 1727203840.11183: variable 'item' from source: include params 13731 1727203840.11204: variable 'ansible_search_path' from source: unknown 13731 1727203840.11205: variable 'ansible_search_path' from source: unknown 13731 1727203840.11208: variable 'item' from source: include params 13731 1727203840.11244: variable 'item' from source: include params 13731 1727203840.11262: we have included files to process 13731 1727203840.11263: generating all_blocks data 13731 1727203840.11264: done generating all_blocks data 13731 1727203840.11267: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.11268: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.11270: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.11391: in VariableManager get_vars() 13731 1727203840.11404: done with get_vars() 13731 1727203840.11570: done processing included file 13731 1727203840.11571: iterating over new_blocks loaded from include file 13731 1727203840.11572: in VariableManager get_vars() 13731 1727203840.11583: done with get_vars() 13731 1727203840.11585: filtering new block on tags 13731 1727203840.11619: done filtering new block on tags 13731 1727203840.11621: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml for managed-node3 => (item=bond0) 13731 1727203840.11625: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.11625: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.11627: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.11687: in VariableManager get_vars() 13731 1727203840.11699: done with get_vars() 13731 1727203840.11878: done processing included file 13731 1727203840.11880: iterating over new_blocks loaded from include file 13731 1727203840.11880: in VariableManager get_vars() 13731 1727203840.11929: done with get_vars() 13731 1727203840.11931: filtering new block on tags 13731 1727203840.11963: done filtering new block on tags 13731 1727203840.11965: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml for managed-node3 => (item=bond0.0) 13731 1727203840.11967: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.11968: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.11970: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml 13731 1727203840.12039: in VariableManager get_vars() 13731 1727203840.12051: done with get_vars() 13731 1727203840.12203: done processing included file 13731 1727203840.12204: iterating over new_blocks loaded from include file 13731 1727203840.12205: in VariableManager get_vars() 13731 1727203840.12213: done with get_vars() 13731 1727203840.12214: filtering new block on tags 13731 1727203840.12243: done filtering new block on tags 13731 1727203840.12245: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml for managed-node3 => (item=bond0.1) 13731 1727203840.12247: extending task lists for all hosts with included blocks 13731 1727203840.12314: done extending task lists 13731 1727203840.12315: done processing included files 13731 1727203840.12315: results queue empty 13731 1727203840.12316: checking for any_errors_fatal 13731 1727203840.12318: done checking for any_errors_fatal 13731 1727203840.12318: checking for max_fail_percentage 13731 1727203840.12319: done checking for max_fail_percentage 13731 1727203840.12319: checking to see if all hosts have failed and the running result is not ok 13731 1727203840.12320: done checking to see if all hosts have failed 13731 1727203840.12320: getting the remaining hosts for this loop 13731 1727203840.12321: done getting the remaining hosts for this loop 13731 1727203840.12322: getting the next task for host managed-node3 13731 1727203840.12325: done getting next task for host managed-node3 13731 1727203840.12327: ^ task is: TASK: Include the task 'get_profile_stat.yml' 13731 1727203840.12329: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203840.12330: getting variables 13731 1727203840.12331: in VariableManager get_vars() 13731 1727203840.12336: Calling all_inventory to load vars for managed-node3 13731 1727203840.12337: Calling groups_inventory to load vars for managed-node3 13731 1727203840.12339: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.12342: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.12344: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.12346: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.13016: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.14053: done with get_vars() 13731 1727203840.14075: done getting variables TASK [Include the task 'get_profile_stat.yml'] ********************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:3 Tuesday 24 September 2024 14:50:40 -0400 (0:00:00.067) 0:00:18.372 ***** 13731 1727203840.14144: entering _queue_task() for managed-node3/include_tasks 13731 1727203840.14460: worker is 1 (out of 1 available) 13731 1727203840.14477: exiting _queue_task() for managed-node3/include_tasks 13731 1727203840.14489: done queuing things up, now waiting for results queue to drain 13731 1727203840.14491: waiting for pending results... 13731 1727203840.14804: running TaskExecutor() for managed-node3/TASK: Include the task 'get_profile_stat.yml' 13731 1727203840.14886: in run() - task 028d2410-947f-82dc-c122-0000000004d9 13731 1727203840.14897: variable 'ansible_search_path' from source: unknown 13731 1727203840.14900: variable 'ansible_search_path' from source: unknown 13731 1727203840.14936: calling self._execute() 13731 1727203840.15010: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.15014: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.15023: variable 'omit' from source: magic vars 13731 1727203840.15298: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.15308: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.15313: _execute() done 13731 1727203840.15316: dumping result to json 13731 1727203840.15318: done dumping result, returning 13731 1727203840.15324: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_profile_stat.yml' [028d2410-947f-82dc-c122-0000000004d9] 13731 1727203840.15329: sending task result for task 028d2410-947f-82dc-c122-0000000004d9 13731 1727203840.15409: done sending task result for task 028d2410-947f-82dc-c122-0000000004d9 13731 1727203840.15412: WORKER PROCESS EXITING 13731 1727203840.15436: no more pending results, returning what we have 13731 1727203840.15441: in VariableManager get_vars() 13731 1727203840.15478: Calling all_inventory to load vars for managed-node3 13731 1727203840.15481: Calling groups_inventory to load vars for managed-node3 13731 1727203840.15484: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.15497: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.15500: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.15502: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.16396: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.18212: done with get_vars() 13731 1727203840.18235: variable 'ansible_search_path' from source: unknown 13731 1727203840.18236: variable 'ansible_search_path' from source: unknown 13731 1727203840.18274: we have included files to process 13731 1727203840.18278: generating all_blocks data 13731 1727203840.18279: done generating all_blocks data 13731 1727203840.18281: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203840.18282: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203840.18284: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203840.19608: done processing included file 13731 1727203840.19611: iterating over new_blocks loaded from include file 13731 1727203840.19612: in VariableManager get_vars() 13731 1727203840.19631: done with get_vars() 13731 1727203840.19633: filtering new block on tags 13731 1727203840.19969: done filtering new block on tags 13731 1727203840.19973: in VariableManager get_vars() 13731 1727203840.19991: done with get_vars() 13731 1727203840.19993: filtering new block on tags 13731 1727203840.20050: done filtering new block on tags 13731 1727203840.20053: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml for managed-node3 13731 1727203840.20058: extending task lists for all hosts with included blocks 13731 1727203840.20473: done extending task lists 13731 1727203840.20477: done processing included files 13731 1727203840.20478: results queue empty 13731 1727203840.20478: checking for any_errors_fatal 13731 1727203840.20482: done checking for any_errors_fatal 13731 1727203840.20483: checking for max_fail_percentage 13731 1727203840.20484: done checking for max_fail_percentage 13731 1727203840.20485: checking to see if all hosts have failed and the running result is not ok 13731 1727203840.20485: done checking to see if all hosts have failed 13731 1727203840.20486: getting the remaining hosts for this loop 13731 1727203840.20487: done getting the remaining hosts for this loop 13731 1727203840.20490: getting the next task for host managed-node3 13731 1727203840.20494: done getting next task for host managed-node3 13731 1727203840.20497: ^ task is: TASK: Initialize NM profile exist and ansible_managed comment flag 13731 1727203840.20501: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203840.20503: getting variables 13731 1727203840.20504: in VariableManager get_vars() 13731 1727203840.20512: Calling all_inventory to load vars for managed-node3 13731 1727203840.20514: Calling groups_inventory to load vars for managed-node3 13731 1727203840.20516: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.20522: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.20524: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.20526: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.21696: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.23232: done with get_vars() 13731 1727203840.23254: done getting variables 13731 1727203840.23298: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Initialize NM profile exist and ansible_managed comment flag] ************ task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:3 Tuesday 24 September 2024 14:50:40 -0400 (0:00:00.091) 0:00:18.464 ***** 13731 1727203840.23329: entering _queue_task() for managed-node3/set_fact 13731 1727203840.23659: worker is 1 (out of 1 available) 13731 1727203840.23670: exiting _queue_task() for managed-node3/set_fact 13731 1727203840.23683: done queuing things up, now waiting for results queue to drain 13731 1727203840.23685: waiting for pending results... 13731 1727203840.24100: running TaskExecutor() for managed-node3/TASK: Initialize NM profile exist and ansible_managed comment flag 13731 1727203840.24116: in run() - task 028d2410-947f-82dc-c122-0000000004fc 13731 1727203840.24138: variable 'ansible_search_path' from source: unknown 13731 1727203840.24145: variable 'ansible_search_path' from source: unknown 13731 1727203840.24189: calling self._execute() 13731 1727203840.24286: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.24303: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.24480: variable 'omit' from source: magic vars 13731 1727203840.24720: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.24736: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.24745: variable 'omit' from source: magic vars 13731 1727203840.24814: variable 'omit' from source: magic vars 13731 1727203840.24856: variable 'omit' from source: magic vars 13731 1727203840.24912: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203840.24955: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203840.24984: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203840.25006: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203840.25040: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203840.25077: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203840.25088: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.25096: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.25202: Set connection var ansible_pipelining to False 13731 1727203840.25213: Set connection var ansible_shell_type to sh 13731 1727203840.25229: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203840.25242: Set connection var ansible_connection to ssh 13731 1727203840.25257: Set connection var ansible_shell_executable to /bin/sh 13731 1727203840.25267: Set connection var ansible_timeout to 10 13731 1727203840.25293: variable 'ansible_shell_executable' from source: unknown 13731 1727203840.25359: variable 'ansible_connection' from source: unknown 13731 1727203840.25363: variable 'ansible_module_compression' from source: unknown 13731 1727203840.25365: variable 'ansible_shell_type' from source: unknown 13731 1727203840.25368: variable 'ansible_shell_executable' from source: unknown 13731 1727203840.25370: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.25372: variable 'ansible_pipelining' from source: unknown 13731 1727203840.25374: variable 'ansible_timeout' from source: unknown 13731 1727203840.25378: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.25490: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203840.25507: variable 'omit' from source: magic vars 13731 1727203840.25518: starting attempt loop 13731 1727203840.25525: running the handler 13731 1727203840.25542: handler run complete 13731 1727203840.25556: attempt loop complete, returning result 13731 1727203840.25576: _execute() done 13731 1727203840.25580: dumping result to json 13731 1727203840.25582: done dumping result, returning 13731 1727203840.25682: done running TaskExecutor() for managed-node3/TASK: Initialize NM profile exist and ansible_managed comment flag [028d2410-947f-82dc-c122-0000000004fc] 13731 1727203840.25686: sending task result for task 028d2410-947f-82dc-c122-0000000004fc 13731 1727203840.25755: done sending task result for task 028d2410-947f-82dc-c122-0000000004fc 13731 1727203840.25759: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "lsr_net_profile_ansible_managed": false, "lsr_net_profile_exists": false, "lsr_net_profile_fingerprint": false }, "changed": false } 13731 1727203840.25841: no more pending results, returning what we have 13731 1727203840.25846: results queue empty 13731 1727203840.25847: checking for any_errors_fatal 13731 1727203840.25848: done checking for any_errors_fatal 13731 1727203840.25849: checking for max_fail_percentage 13731 1727203840.25851: done checking for max_fail_percentage 13731 1727203840.25852: checking to see if all hosts have failed and the running result is not ok 13731 1727203840.25853: done checking to see if all hosts have failed 13731 1727203840.25853: getting the remaining hosts for this loop 13731 1727203840.25856: done getting the remaining hosts for this loop 13731 1727203840.25860: getting the next task for host managed-node3 13731 1727203840.25870: done getting next task for host managed-node3 13731 1727203840.25873: ^ task is: TASK: Stat profile file 13731 1727203840.25883: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203840.25888: getting variables 13731 1727203840.25890: in VariableManager get_vars() 13731 1727203840.25922: Calling all_inventory to load vars for managed-node3 13731 1727203840.25925: Calling groups_inventory to load vars for managed-node3 13731 1727203840.25929: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.25939: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.25942: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.25946: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.28427: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.30608: done with get_vars() 13731 1727203840.30630: done getting variables TASK [Stat profile file] ******************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:9 Tuesday 24 September 2024 14:50:40 -0400 (0:00:00.074) 0:00:18.538 ***** 13731 1727203840.30744: entering _queue_task() for managed-node3/stat 13731 1727203840.31100: worker is 1 (out of 1 available) 13731 1727203840.31113: exiting _queue_task() for managed-node3/stat 13731 1727203840.31237: done queuing things up, now waiting for results queue to drain 13731 1727203840.31239: waiting for pending results... 13731 1727203840.31445: running TaskExecutor() for managed-node3/TASK: Stat profile file 13731 1727203840.31603: in run() - task 028d2410-947f-82dc-c122-0000000004fd 13731 1727203840.31612: variable 'ansible_search_path' from source: unknown 13731 1727203840.31617: variable 'ansible_search_path' from source: unknown 13731 1727203840.31666: calling self._execute() 13731 1727203840.31796: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.31805: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.31813: variable 'omit' from source: magic vars 13731 1727203840.32097: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.32108: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.32113: variable 'omit' from source: magic vars 13731 1727203840.32152: variable 'omit' from source: magic vars 13731 1727203840.32224: variable 'profile' from source: include params 13731 1727203840.32228: variable 'bond_port_profile' from source: include params 13731 1727203840.32277: variable 'bond_port_profile' from source: include params 13731 1727203840.32291: variable 'omit' from source: magic vars 13731 1727203840.32325: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203840.32351: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203840.32370: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203840.32384: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203840.32395: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203840.32419: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203840.32422: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.32424: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.32498: Set connection var ansible_pipelining to False 13731 1727203840.32501: Set connection var ansible_shell_type to sh 13731 1727203840.32507: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203840.32515: Set connection var ansible_connection to ssh 13731 1727203840.32518: Set connection var ansible_shell_executable to /bin/sh 13731 1727203840.32522: Set connection var ansible_timeout to 10 13731 1727203840.32540: variable 'ansible_shell_executable' from source: unknown 13731 1727203840.32543: variable 'ansible_connection' from source: unknown 13731 1727203840.32545: variable 'ansible_module_compression' from source: unknown 13731 1727203840.32547: variable 'ansible_shell_type' from source: unknown 13731 1727203840.32550: variable 'ansible_shell_executable' from source: unknown 13731 1727203840.32552: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.32554: variable 'ansible_pipelining' from source: unknown 13731 1727203840.32558: variable 'ansible_timeout' from source: unknown 13731 1727203840.32560: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.32710: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203840.32718: variable 'omit' from source: magic vars 13731 1727203840.32723: starting attempt loop 13731 1727203840.32727: running the handler 13731 1727203840.32740: _low_level_execute_command(): starting 13731 1727203840.32747: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203840.33242: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.33246: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.33250: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.33293: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.33297: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.33356: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.35025: stdout chunk (state=3): >>>/root <<< 13731 1727203840.35133: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.35141: stdout chunk (state=3): >>><<< 13731 1727203840.35147: stderr chunk (state=3): >>><<< 13731 1727203840.35172: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203840.35181: _low_level_execute_command(): starting 13731 1727203840.35188: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661 `" && echo ansible-tmp-1727203840.3516557-14833-134723266954661="` echo /root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661 `" ) && sleep 0' 13731 1727203840.35628: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.35631: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.35641: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.35643: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.35689: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203840.35693: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.35732: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.37581: stdout chunk (state=3): >>>ansible-tmp-1727203840.3516557-14833-134723266954661=/root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661 <<< 13731 1727203840.37706: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.37718: stderr chunk (state=3): >>><<< 13731 1727203840.37721: stdout chunk (state=3): >>><<< 13731 1727203840.37741: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203840.3516557-14833-134723266954661=/root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203840.37783: variable 'ansible_module_compression' from source: unknown 13731 1727203840.37846: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13731 1727203840.37994: variable 'ansible_facts' from source: unknown 13731 1727203840.37997: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/AnsiballZ_stat.py 13731 1727203840.38288: Sending initial data 13731 1727203840.38292: Sent initial data (153 bytes) 13731 1727203840.38632: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203840.38641: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203840.38652: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.38746: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.38764: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.38817: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.40315: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203840.40343: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203840.40382: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp2ug59vzf /root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/AnsiballZ_stat.py <<< 13731 1727203840.40386: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/AnsiballZ_stat.py" <<< 13731 1727203840.40440: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp2ug59vzf" to remote "/root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/AnsiballZ_stat.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/AnsiballZ_stat.py" <<< 13731 1727203840.40927: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.41109: stderr chunk (state=3): >>><<< 13731 1727203840.41112: stdout chunk (state=3): >>><<< 13731 1727203840.41115: done transferring module to remote 13731 1727203840.41117: _low_level_execute_command(): starting 13731 1727203840.41119: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/ /root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/AnsiballZ_stat.py && sleep 0' 13731 1727203840.41646: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203840.41664: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203840.41681: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.41775: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.41812: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203840.41825: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.41847: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.41913: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.43632: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.43644: stdout chunk (state=3): >>><<< 13731 1727203840.43650: stderr chunk (state=3): >>><<< 13731 1727203840.43664: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203840.43667: _low_level_execute_command(): starting 13731 1727203840.43670: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/AnsiballZ_stat.py && sleep 0' 13731 1727203840.44066: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.44070: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203840.44103: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.44106: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.44108: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.44149: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.44165: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.44206: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.59484: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13731 1727203840.60778: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203840.60807: stderr chunk (state=3): >>><<< 13731 1727203840.60810: stdout chunk (state=3): >>><<< 13731 1727203840.60826: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203840.60853: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/etc/sysconfig/network-scripts/ifcfg-bond0', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203840.60859: _low_level_execute_command(): starting 13731 1727203840.60865: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203840.3516557-14833-134723266954661/ > /dev/null 2>&1 && sleep 0' 13731 1727203840.61492: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.61536: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203840.61554: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.61584: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.61646: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.63472: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.63505: stderr chunk (state=3): >>><<< 13731 1727203840.63508: stdout chunk (state=3): >>><<< 13731 1727203840.63523: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203840.63528: handler run complete 13731 1727203840.63544: attempt loop complete, returning result 13731 1727203840.63547: _execute() done 13731 1727203840.63549: dumping result to json 13731 1727203840.63552: done dumping result, returning 13731 1727203840.63560: done running TaskExecutor() for managed-node3/TASK: Stat profile file [028d2410-947f-82dc-c122-0000000004fd] 13731 1727203840.63565: sending task result for task 028d2410-947f-82dc-c122-0000000004fd 13731 1727203840.63653: done sending task result for task 028d2410-947f-82dc-c122-0000000004fd 13731 1727203840.63656: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } 13731 1727203840.63749: no more pending results, returning what we have 13731 1727203840.63753: results queue empty 13731 1727203840.63754: checking for any_errors_fatal 13731 1727203840.63766: done checking for any_errors_fatal 13731 1727203840.63767: checking for max_fail_percentage 13731 1727203840.63768: done checking for max_fail_percentage 13731 1727203840.63769: checking to see if all hosts have failed and the running result is not ok 13731 1727203840.63770: done checking to see if all hosts have failed 13731 1727203840.63770: getting the remaining hosts for this loop 13731 1727203840.63772: done getting the remaining hosts for this loop 13731 1727203840.63777: getting the next task for host managed-node3 13731 1727203840.63785: done getting next task for host managed-node3 13731 1727203840.63787: ^ task is: TASK: Set NM profile exist flag based on the profile files 13731 1727203840.63792: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203840.63796: getting variables 13731 1727203840.63797: in VariableManager get_vars() 13731 1727203840.63824: Calling all_inventory to load vars for managed-node3 13731 1727203840.63826: Calling groups_inventory to load vars for managed-node3 13731 1727203840.63829: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.63839: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.63841: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.63843: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.65498: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.67053: done with get_vars() 13731 1727203840.67080: done getting variables 13731 1727203840.67140: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag based on the profile files] ******************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:17 Tuesday 24 September 2024 14:50:40 -0400 (0:00:00.364) 0:00:18.902 ***** 13731 1727203840.67173: entering _queue_task() for managed-node3/set_fact 13731 1727203840.67597: worker is 1 (out of 1 available) 13731 1727203840.67608: exiting _queue_task() for managed-node3/set_fact 13731 1727203840.67618: done queuing things up, now waiting for results queue to drain 13731 1727203840.67619: waiting for pending results... 13731 1727203840.67905: running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag based on the profile files 13731 1727203840.67933: in run() - task 028d2410-947f-82dc-c122-0000000004fe 13731 1727203840.67981: variable 'ansible_search_path' from source: unknown 13731 1727203840.67985: variable 'ansible_search_path' from source: unknown 13731 1727203840.67990: calling self._execute() 13731 1727203840.68085: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.68090: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.68098: variable 'omit' from source: magic vars 13731 1727203840.68512: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.68515: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.68607: variable 'profile_stat' from source: set_fact 13731 1727203840.68624: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203840.68628: when evaluation is False, skipping this task 13731 1727203840.68630: _execute() done 13731 1727203840.68633: dumping result to json 13731 1727203840.68635: done dumping result, returning 13731 1727203840.68639: done running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag based on the profile files [028d2410-947f-82dc-c122-0000000004fe] 13731 1727203840.68646: sending task result for task 028d2410-947f-82dc-c122-0000000004fe 13731 1727203840.68790: done sending task result for task 028d2410-947f-82dc-c122-0000000004fe 13731 1727203840.68794: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203840.68916: no more pending results, returning what we have 13731 1727203840.68920: results queue empty 13731 1727203840.68921: checking for any_errors_fatal 13731 1727203840.68928: done checking for any_errors_fatal 13731 1727203840.68929: checking for max_fail_percentage 13731 1727203840.68930: done checking for max_fail_percentage 13731 1727203840.68931: checking to see if all hosts have failed and the running result is not ok 13731 1727203840.68932: done checking to see if all hosts have failed 13731 1727203840.68932: getting the remaining hosts for this loop 13731 1727203840.68934: done getting the remaining hosts for this loop 13731 1727203840.68937: getting the next task for host managed-node3 13731 1727203840.68944: done getting next task for host managed-node3 13731 1727203840.68946: ^ task is: TASK: Get NM profile info 13731 1727203840.68952: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203840.68956: getting variables 13731 1727203840.68957: in VariableManager get_vars() 13731 1727203840.68988: Calling all_inventory to load vars for managed-node3 13731 1727203840.68992: Calling groups_inventory to load vars for managed-node3 13731 1727203840.68995: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203840.69006: Calling all_plugins_play to load vars for managed-node3 13731 1727203840.69009: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203840.69011: Calling groups_plugins_play to load vars for managed-node3 13731 1727203840.70528: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203840.71758: done with get_vars() 13731 1727203840.71774: done getting variables 13731 1727203840.71819: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Get NM profile info] ***************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:25 Tuesday 24 September 2024 14:50:40 -0400 (0:00:00.046) 0:00:18.949 ***** 13731 1727203840.71841: entering _queue_task() for managed-node3/shell 13731 1727203840.72064: worker is 1 (out of 1 available) 13731 1727203840.72079: exiting _queue_task() for managed-node3/shell 13731 1727203840.72091: done queuing things up, now waiting for results queue to drain 13731 1727203840.72093: waiting for pending results... 13731 1727203840.72279: running TaskExecutor() for managed-node3/TASK: Get NM profile info 13731 1727203840.72354: in run() - task 028d2410-947f-82dc-c122-0000000004ff 13731 1727203840.72367: variable 'ansible_search_path' from source: unknown 13731 1727203840.72370: variable 'ansible_search_path' from source: unknown 13731 1727203840.72399: calling self._execute() 13731 1727203840.72467: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.72473: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.72483: variable 'omit' from source: magic vars 13731 1727203840.72763: variable 'ansible_distribution_major_version' from source: facts 13731 1727203840.72770: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203840.72773: variable 'omit' from source: magic vars 13731 1727203840.72811: variable 'omit' from source: magic vars 13731 1727203840.72881: variable 'profile' from source: include params 13731 1727203840.72885: variable 'bond_port_profile' from source: include params 13731 1727203840.72933: variable 'bond_port_profile' from source: include params 13731 1727203840.72947: variable 'omit' from source: magic vars 13731 1727203840.72983: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203840.73009: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203840.73025: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203840.73037: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203840.73047: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203840.73075: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203840.73079: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.73083: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.73149: Set connection var ansible_pipelining to False 13731 1727203840.73152: Set connection var ansible_shell_type to sh 13731 1727203840.73158: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203840.73165: Set connection var ansible_connection to ssh 13731 1727203840.73168: Set connection var ansible_shell_executable to /bin/sh 13731 1727203840.73173: Set connection var ansible_timeout to 10 13731 1727203840.73190: variable 'ansible_shell_executable' from source: unknown 13731 1727203840.73199: variable 'ansible_connection' from source: unknown 13731 1727203840.73203: variable 'ansible_module_compression' from source: unknown 13731 1727203840.73206: variable 'ansible_shell_type' from source: unknown 13731 1727203840.73208: variable 'ansible_shell_executable' from source: unknown 13731 1727203840.73210: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203840.73212: variable 'ansible_pipelining' from source: unknown 13731 1727203840.73214: variable 'ansible_timeout' from source: unknown 13731 1727203840.73216: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203840.73405: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203840.73411: variable 'omit' from source: magic vars 13731 1727203840.73413: starting attempt loop 13731 1727203840.73416: running the handler 13731 1727203840.73418: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203840.73420: _low_level_execute_command(): starting 13731 1727203840.73422: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203840.74155: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.74163: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203840.74167: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.74169: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.74218: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.75797: stdout chunk (state=3): >>>/root <<< 13731 1727203840.75899: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.75929: stderr chunk (state=3): >>><<< 13731 1727203840.75931: stdout chunk (state=3): >>><<< 13731 1727203840.75945: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203840.75959: _low_level_execute_command(): starting 13731 1727203840.75984: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812 `" && echo ansible-tmp-1727203840.7595036-14854-177196037521812="` echo /root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812 `" ) && sleep 0' 13731 1727203840.76455: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.76466: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.76470: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203840.76472: stderr chunk (state=3): >>>debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.76508: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203840.76512: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.76558: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.78412: stdout chunk (state=3): >>>ansible-tmp-1727203840.7595036-14854-177196037521812=/root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812 <<< 13731 1727203840.78556: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.78560: stdout chunk (state=3): >>><<< 13731 1727203840.78562: stderr chunk (state=3): >>><<< 13731 1727203840.78580: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203840.7595036-14854-177196037521812=/root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203840.78609: variable 'ansible_module_compression' from source: unknown 13731 1727203840.78662: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203840.78696: variable 'ansible_facts' from source: unknown 13731 1727203840.78749: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/AnsiballZ_command.py 13731 1727203840.78852: Sending initial data 13731 1727203840.78856: Sent initial data (156 bytes) 13731 1727203840.79282: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.79285: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.79288: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.79290: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.79370: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.79373: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.79411: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.81136: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 debug2: Sending SSH2_FXP_REALPATH "." debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/AnsiballZ_command.py" <<< 13731 1727203840.81140: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp8si0cnm2 /root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/AnsiballZ_command.py <<< 13731 1727203840.81316: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp8si0cnm2" to remote "/root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/AnsiballZ_command.py" <<< 13731 1727203840.82215: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.82284: stderr chunk (state=3): >>><<< 13731 1727203840.82297: stdout chunk (state=3): >>><<< 13731 1727203840.82326: done transferring module to remote 13731 1727203840.82339: _low_level_execute_command(): starting 13731 1727203840.82342: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/ /root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/AnsiballZ_command.py && sleep 0' 13731 1727203840.82980: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203840.82990: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203840.83001: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203840.83014: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203840.83068: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203840.83124: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203840.83179: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.83190: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.83217: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203840.84970: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203840.84974: stdout chunk (state=3): >>><<< 13731 1727203840.84978: stderr chunk (state=3): >>><<< 13731 1727203840.84994: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203840.85078: _low_level_execute_command(): starting 13731 1727203840.85081: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/AnsiballZ_command.py && sleep 0' 13731 1727203840.85650: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203840.85666: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203840.85737: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203841.02637: stdout chunk (state=3): >>> {"changed": true, "stdout": "bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection \nbond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection \nbond0 /etc/NetworkManager/system-connections/bond0.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "start": "2024-09-24 14:50:41.005761", "end": "2024-09-24 14:50:41.025647", "delta": "0:00:00.019886", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203841.04083: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203841.04087: stderr chunk (state=3): >>><<< 13731 1727203841.04089: stdout chunk (state=3): >>><<< 13731 1727203841.04111: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection \nbond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection \nbond0 /etc/NetworkManager/system-connections/bond0.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "start": "2024-09-24 14:50:41.005761", "end": "2024-09-24 14:50:41.025647", "delta": "0:00:00.019886", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203841.04239: done with _execute_module (ansible.legacy.command, {'_raw_params': 'nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203841.04243: _low_level_execute_command(): starting 13731 1727203841.04245: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203840.7595036-14854-177196037521812/ > /dev/null 2>&1 && sleep 0' 13731 1727203841.04824: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203841.04839: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203841.04855: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203841.04882: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203841.04900: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203841.04995: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203841.05017: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203841.05032: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203841.05100: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203841.06880: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203841.06890: stdout chunk (state=3): >>><<< 13731 1727203841.06901: stderr chunk (state=3): >>><<< 13731 1727203841.06923: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203841.06934: handler run complete 13731 1727203841.06959: Evaluated conditional (False): False 13731 1727203841.06977: attempt loop complete, returning result 13731 1727203841.06985: _execute() done 13731 1727203841.06990: dumping result to json 13731 1727203841.07016: done dumping result, returning 13731 1727203841.07019: done running TaskExecutor() for managed-node3/TASK: Get NM profile info [028d2410-947f-82dc-c122-0000000004ff] 13731 1727203841.07021: sending task result for task 028d2410-947f-82dc-c122-0000000004ff 13731 1727203841.07397: done sending task result for task 028d2410-947f-82dc-c122-0000000004ff 13731 1727203841.07402: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0 | grep /etc", "delta": "0:00:00.019886", "end": "2024-09-24 14:50:41.025647", "rc": 0, "start": "2024-09-24 14:50:41.005761" } STDOUT: bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection bond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection bond0 /etc/NetworkManager/system-connections/bond0.nmconnection 13731 1727203841.07490: no more pending results, returning what we have 13731 1727203841.07494: results queue empty 13731 1727203841.07495: checking for any_errors_fatal 13731 1727203841.07501: done checking for any_errors_fatal 13731 1727203841.07501: checking for max_fail_percentage 13731 1727203841.07503: done checking for max_fail_percentage 13731 1727203841.07504: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.07504: done checking to see if all hosts have failed 13731 1727203841.07505: getting the remaining hosts for this loop 13731 1727203841.07507: done getting the remaining hosts for this loop 13731 1727203841.07510: getting the next task for host managed-node3 13731 1727203841.07517: done getting next task for host managed-node3 13731 1727203841.07520: ^ task is: TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13731 1727203841.07525: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.07528: getting variables 13731 1727203841.07530: in VariableManager get_vars() 13731 1727203841.07564: Calling all_inventory to load vars for managed-node3 13731 1727203841.07567: Calling groups_inventory to load vars for managed-node3 13731 1727203841.07571: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.07684: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.07688: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.07692: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.09039: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.10641: done with get_vars() 13731 1727203841.10673: done getting variables 13731 1727203841.10738: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag and ansible_managed flag true based on the nmcli output] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:35 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.389) 0:00:19.338 ***** 13731 1727203841.10779: entering _queue_task() for managed-node3/set_fact 13731 1727203841.11128: worker is 1 (out of 1 available) 13731 1727203841.11142: exiting _queue_task() for managed-node3/set_fact 13731 1727203841.11154: done queuing things up, now waiting for results queue to drain 13731 1727203841.11155: waiting for pending results... 13731 1727203841.11463: running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13731 1727203841.11608: in run() - task 028d2410-947f-82dc-c122-000000000500 13731 1727203841.11629: variable 'ansible_search_path' from source: unknown 13731 1727203841.11637: variable 'ansible_search_path' from source: unknown 13731 1727203841.11683: calling self._execute() 13731 1727203841.11789: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.11802: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.11822: variable 'omit' from source: magic vars 13731 1727203841.12194: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.12210: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.12342: variable 'nm_profile_exists' from source: set_fact 13731 1727203841.12364: Evaluated conditional (nm_profile_exists.rc == 0): True 13731 1727203841.12377: variable 'omit' from source: magic vars 13731 1727203841.12444: variable 'omit' from source: magic vars 13731 1727203841.12489: variable 'omit' from source: magic vars 13731 1727203841.12532: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203841.12577: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203841.12603: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203841.12624: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.12639: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.12677: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203841.12687: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.12780: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.12805: Set connection var ansible_pipelining to False 13731 1727203841.12815: Set connection var ansible_shell_type to sh 13731 1727203841.12824: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203841.12832: Set connection var ansible_connection to ssh 13731 1727203841.12841: Set connection var ansible_shell_executable to /bin/sh 13731 1727203841.12849: Set connection var ansible_timeout to 10 13731 1727203841.12878: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.12886: variable 'ansible_connection' from source: unknown 13731 1727203841.12893: variable 'ansible_module_compression' from source: unknown 13731 1727203841.12900: variable 'ansible_shell_type' from source: unknown 13731 1727203841.12912: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.12921: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.12930: variable 'ansible_pipelining' from source: unknown 13731 1727203841.12941: variable 'ansible_timeout' from source: unknown 13731 1727203841.12951: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.13108: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203841.13124: variable 'omit' from source: magic vars 13731 1727203841.13238: starting attempt loop 13731 1727203841.13241: running the handler 13731 1727203841.13243: handler run complete 13731 1727203841.13245: attempt loop complete, returning result 13731 1727203841.13248: _execute() done 13731 1727203841.13250: dumping result to json 13731 1727203841.13251: done dumping result, returning 13731 1727203841.13254: done running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output [028d2410-947f-82dc-c122-000000000500] 13731 1727203841.13256: sending task result for task 028d2410-947f-82dc-c122-000000000500 13731 1727203841.13326: done sending task result for task 028d2410-947f-82dc-c122-000000000500 13731 1727203841.13329: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "lsr_net_profile_ansible_managed": true, "lsr_net_profile_exists": true, "lsr_net_profile_fingerprint": true }, "changed": false } 13731 1727203841.13390: no more pending results, returning what we have 13731 1727203841.13393: results queue empty 13731 1727203841.13394: checking for any_errors_fatal 13731 1727203841.13402: done checking for any_errors_fatal 13731 1727203841.13403: checking for max_fail_percentage 13731 1727203841.13404: done checking for max_fail_percentage 13731 1727203841.13405: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.13406: done checking to see if all hosts have failed 13731 1727203841.13407: getting the remaining hosts for this loop 13731 1727203841.13409: done getting the remaining hosts for this loop 13731 1727203841.13413: getting the next task for host managed-node3 13731 1727203841.13423: done getting next task for host managed-node3 13731 1727203841.13426: ^ task is: TASK: Get the ansible_managed comment in ifcfg-{{ profile }} 13731 1727203841.13432: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.13436: getting variables 13731 1727203841.13438: in VariableManager get_vars() 13731 1727203841.13473: Calling all_inventory to load vars for managed-node3 13731 1727203841.13478: Calling groups_inventory to load vars for managed-node3 13731 1727203841.13483: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.13493: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.13496: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.13499: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.15271: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.16752: done with get_vars() 13731 1727203841.16781: done getting variables 13731 1727203841.16832: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203841.16947: variable 'profile' from source: include params 13731 1727203841.16951: variable 'bond_port_profile' from source: include params 13731 1727203841.17014: variable 'bond_port_profile' from source: include params TASK [Get the ansible_managed comment in ifcfg-bond0] ************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:49 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.062) 0:00:19.401 ***** 13731 1727203841.17051: entering _queue_task() for managed-node3/command 13731 1727203841.17354: worker is 1 (out of 1 available) 13731 1727203841.17369: exiting _queue_task() for managed-node3/command 13731 1727203841.17582: done queuing things up, now waiting for results queue to drain 13731 1727203841.17584: waiting for pending results... 13731 1727203841.17713: running TaskExecutor() for managed-node3/TASK: Get the ansible_managed comment in ifcfg-bond0 13731 1727203841.17792: in run() - task 028d2410-947f-82dc-c122-000000000502 13731 1727203841.17917: variable 'ansible_search_path' from source: unknown 13731 1727203841.17920: variable 'ansible_search_path' from source: unknown 13731 1727203841.17923: calling self._execute() 13731 1727203841.17953: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.17965: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.17977: variable 'omit' from source: magic vars 13731 1727203841.18331: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.18349: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.18466: variable 'profile_stat' from source: set_fact 13731 1727203841.18483: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203841.18492: when evaluation is False, skipping this task 13731 1727203841.18499: _execute() done 13731 1727203841.18506: dumping result to json 13731 1727203841.18512: done dumping result, returning 13731 1727203841.18521: done running TaskExecutor() for managed-node3/TASK: Get the ansible_managed comment in ifcfg-bond0 [028d2410-947f-82dc-c122-000000000502] 13731 1727203841.18532: sending task result for task 028d2410-947f-82dc-c122-000000000502 13731 1727203841.18782: done sending task result for task 028d2410-947f-82dc-c122-000000000502 13731 1727203841.18785: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203841.18828: no more pending results, returning what we have 13731 1727203841.18831: results queue empty 13731 1727203841.18832: checking for any_errors_fatal 13731 1727203841.18839: done checking for any_errors_fatal 13731 1727203841.18840: checking for max_fail_percentage 13731 1727203841.18841: done checking for max_fail_percentage 13731 1727203841.18842: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.18843: done checking to see if all hosts have failed 13731 1727203841.18844: getting the remaining hosts for this loop 13731 1727203841.18845: done getting the remaining hosts for this loop 13731 1727203841.18849: getting the next task for host managed-node3 13731 1727203841.18855: done getting next task for host managed-node3 13731 1727203841.18858: ^ task is: TASK: Verify the ansible_managed comment in ifcfg-{{ profile }} 13731 1727203841.18866: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.18870: getting variables 13731 1727203841.18871: in VariableManager get_vars() 13731 1727203841.18903: Calling all_inventory to load vars for managed-node3 13731 1727203841.18906: Calling groups_inventory to load vars for managed-node3 13731 1727203841.18910: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.18921: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.18924: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.18927: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.20338: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.21931: done with get_vars() 13731 1727203841.21955: done getting variables 13731 1727203841.22013: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203841.22116: variable 'profile' from source: include params 13731 1727203841.22119: variable 'bond_port_profile' from source: include params 13731 1727203841.22172: variable 'bond_port_profile' from source: include params TASK [Verify the ansible_managed comment in ifcfg-bond0] *********************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:56 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.051) 0:00:19.453 ***** 13731 1727203841.22209: entering _queue_task() for managed-node3/set_fact 13731 1727203841.22541: worker is 1 (out of 1 available) 13731 1727203841.22554: exiting _queue_task() for managed-node3/set_fact 13731 1727203841.22569: done queuing things up, now waiting for results queue to drain 13731 1727203841.22571: waiting for pending results... 13731 1727203841.22871: running TaskExecutor() for managed-node3/TASK: Verify the ansible_managed comment in ifcfg-bond0 13731 1727203841.23025: in run() - task 028d2410-947f-82dc-c122-000000000503 13731 1727203841.23047: variable 'ansible_search_path' from source: unknown 13731 1727203841.23057: variable 'ansible_search_path' from source: unknown 13731 1727203841.23106: calling self._execute() 13731 1727203841.23381: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.23384: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.23387: variable 'omit' from source: magic vars 13731 1727203841.23596: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.23619: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.23742: variable 'profile_stat' from source: set_fact 13731 1727203841.23759: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203841.23767: when evaluation is False, skipping this task 13731 1727203841.23776: _execute() done 13731 1727203841.23784: dumping result to json 13731 1727203841.23792: done dumping result, returning 13731 1727203841.23801: done running TaskExecutor() for managed-node3/TASK: Verify the ansible_managed comment in ifcfg-bond0 [028d2410-947f-82dc-c122-000000000503] 13731 1727203841.23812: sending task result for task 028d2410-947f-82dc-c122-000000000503 skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203841.23982: no more pending results, returning what we have 13731 1727203841.23988: results queue empty 13731 1727203841.23989: checking for any_errors_fatal 13731 1727203841.23997: done checking for any_errors_fatal 13731 1727203841.23998: checking for max_fail_percentage 13731 1727203841.23999: done checking for max_fail_percentage 13731 1727203841.24000: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.24001: done checking to see if all hosts have failed 13731 1727203841.24002: getting the remaining hosts for this loop 13731 1727203841.24004: done getting the remaining hosts for this loop 13731 1727203841.24007: getting the next task for host managed-node3 13731 1727203841.24016: done getting next task for host managed-node3 13731 1727203841.24019: ^ task is: TASK: Get the fingerprint comment in ifcfg-{{ profile }} 13731 1727203841.24026: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.24030: getting variables 13731 1727203841.24032: in VariableManager get_vars() 13731 1727203841.24064: Calling all_inventory to load vars for managed-node3 13731 1727203841.24067: Calling groups_inventory to load vars for managed-node3 13731 1727203841.24072: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.24188: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.24192: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.24195: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.24886: done sending task result for task 028d2410-947f-82dc-c122-000000000503 13731 1727203841.24889: WORKER PROCESS EXITING 13731 1727203841.25732: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.27206: done with get_vars() 13731 1727203841.27226: done getting variables 13731 1727203841.27285: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203841.27394: variable 'profile' from source: include params 13731 1727203841.27397: variable 'bond_port_profile' from source: include params 13731 1727203841.27453: variable 'bond_port_profile' from source: include params TASK [Get the fingerprint comment in ifcfg-bond0] ****************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:62 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.052) 0:00:19.505 ***** 13731 1727203841.27484: entering _queue_task() for managed-node3/command 13731 1727203841.27789: worker is 1 (out of 1 available) 13731 1727203841.27801: exiting _queue_task() for managed-node3/command 13731 1727203841.27811: done queuing things up, now waiting for results queue to drain 13731 1727203841.27812: waiting for pending results... 13731 1727203841.28071: running TaskExecutor() for managed-node3/TASK: Get the fingerprint comment in ifcfg-bond0 13731 1727203841.28204: in run() - task 028d2410-947f-82dc-c122-000000000504 13731 1727203841.28222: variable 'ansible_search_path' from source: unknown 13731 1727203841.28229: variable 'ansible_search_path' from source: unknown 13731 1727203841.28266: calling self._execute() 13731 1727203841.28364: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.28380: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.28396: variable 'omit' from source: magic vars 13731 1727203841.28762: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.28781: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.28905: variable 'profile_stat' from source: set_fact 13731 1727203841.28923: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203841.28930: when evaluation is False, skipping this task 13731 1727203841.28937: _execute() done 13731 1727203841.28945: dumping result to json 13731 1727203841.28954: done dumping result, returning 13731 1727203841.28962: done running TaskExecutor() for managed-node3/TASK: Get the fingerprint comment in ifcfg-bond0 [028d2410-947f-82dc-c122-000000000504] 13731 1727203841.28972: sending task result for task 028d2410-947f-82dc-c122-000000000504 skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203841.29111: no more pending results, returning what we have 13731 1727203841.29116: results queue empty 13731 1727203841.29117: checking for any_errors_fatal 13731 1727203841.29126: done checking for any_errors_fatal 13731 1727203841.29127: checking for max_fail_percentage 13731 1727203841.29129: done checking for max_fail_percentage 13731 1727203841.29130: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.29130: done checking to see if all hosts have failed 13731 1727203841.29131: getting the remaining hosts for this loop 13731 1727203841.29133: done getting the remaining hosts for this loop 13731 1727203841.29136: getting the next task for host managed-node3 13731 1727203841.29146: done getting next task for host managed-node3 13731 1727203841.29148: ^ task is: TASK: Verify the fingerprint comment in ifcfg-{{ profile }} 13731 1727203841.29154: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.29159: getting variables 13731 1727203841.29160: in VariableManager get_vars() 13731 1727203841.29192: Calling all_inventory to load vars for managed-node3 13731 1727203841.29195: Calling groups_inventory to load vars for managed-node3 13731 1727203841.29199: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.29212: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.29215: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.29218: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.29991: done sending task result for task 028d2410-947f-82dc-c122-000000000504 13731 1727203841.29995: WORKER PROCESS EXITING 13731 1727203841.30755: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.32385: done with get_vars() 13731 1727203841.32407: done getting variables 13731 1727203841.32472: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203841.32582: variable 'profile' from source: include params 13731 1727203841.32586: variable 'bond_port_profile' from source: include params 13731 1727203841.32647: variable 'bond_port_profile' from source: include params TASK [Verify the fingerprint comment in ifcfg-bond0] *************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:69 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.051) 0:00:19.557 ***** 13731 1727203841.32683: entering _queue_task() for managed-node3/set_fact 13731 1727203841.33108: worker is 1 (out of 1 available) 13731 1727203841.33121: exiting _queue_task() for managed-node3/set_fact 13731 1727203841.33132: done queuing things up, now waiting for results queue to drain 13731 1727203841.33133: waiting for pending results... 13731 1727203841.33322: running TaskExecutor() for managed-node3/TASK: Verify the fingerprint comment in ifcfg-bond0 13731 1727203841.33464: in run() - task 028d2410-947f-82dc-c122-000000000505 13731 1727203841.33681: variable 'ansible_search_path' from source: unknown 13731 1727203841.33685: variable 'ansible_search_path' from source: unknown 13731 1727203841.33688: calling self._execute() 13731 1727203841.33691: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.33694: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.33696: variable 'omit' from source: magic vars 13731 1727203841.33992: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.34007: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.34123: variable 'profile_stat' from source: set_fact 13731 1727203841.34143: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203841.34150: when evaluation is False, skipping this task 13731 1727203841.34157: _execute() done 13731 1727203841.34163: dumping result to json 13731 1727203841.34169: done dumping result, returning 13731 1727203841.34180: done running TaskExecutor() for managed-node3/TASK: Verify the fingerprint comment in ifcfg-bond0 [028d2410-947f-82dc-c122-000000000505] 13731 1727203841.34189: sending task result for task 028d2410-947f-82dc-c122-000000000505 skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203841.34324: no more pending results, returning what we have 13731 1727203841.34328: results queue empty 13731 1727203841.34329: checking for any_errors_fatal 13731 1727203841.34336: done checking for any_errors_fatal 13731 1727203841.34337: checking for max_fail_percentage 13731 1727203841.34338: done checking for max_fail_percentage 13731 1727203841.34339: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.34340: done checking to see if all hosts have failed 13731 1727203841.34341: getting the remaining hosts for this loop 13731 1727203841.34343: done getting the remaining hosts for this loop 13731 1727203841.34346: getting the next task for host managed-node3 13731 1727203841.34357: done getting next task for host managed-node3 13731 1727203841.34360: ^ task is: TASK: Assert that the profile is present - '{{ profile }}' 13731 1727203841.34365: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.34370: getting variables 13731 1727203841.34371: in VariableManager get_vars() 13731 1727203841.34406: Calling all_inventory to load vars for managed-node3 13731 1727203841.34410: Calling groups_inventory to load vars for managed-node3 13731 1727203841.34414: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.34428: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.34431: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.34434: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.35095: done sending task result for task 028d2410-947f-82dc-c122-000000000505 13731 1727203841.35099: WORKER PROCESS EXITING 13731 1727203841.36025: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.37555: done with get_vars() 13731 1727203841.37583: done getting variables 13731 1727203841.37646: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203841.37769: variable 'profile' from source: include params 13731 1727203841.37773: variable 'bond_port_profile' from source: include params 13731 1727203841.37836: variable 'bond_port_profile' from source: include params TASK [Assert that the profile is present - 'bond0'] **************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:5 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.051) 0:00:19.609 ***** 13731 1727203841.37871: entering _queue_task() for managed-node3/assert 13731 1727203841.38300: worker is 1 (out of 1 available) 13731 1727203841.38311: exiting _queue_task() for managed-node3/assert 13731 1727203841.38322: done queuing things up, now waiting for results queue to drain 13731 1727203841.38324: waiting for pending results... 13731 1727203841.38521: running TaskExecutor() for managed-node3/TASK: Assert that the profile is present - 'bond0' 13731 1727203841.38653: in run() - task 028d2410-947f-82dc-c122-0000000004da 13731 1727203841.38680: variable 'ansible_search_path' from source: unknown 13731 1727203841.38689: variable 'ansible_search_path' from source: unknown 13731 1727203841.38729: calling self._execute() 13731 1727203841.38834: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.38847: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.38864: variable 'omit' from source: magic vars 13731 1727203841.39236: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.39253: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.39264: variable 'omit' from source: magic vars 13731 1727203841.39328: variable 'omit' from source: magic vars 13731 1727203841.39431: variable 'profile' from source: include params 13731 1727203841.39440: variable 'bond_port_profile' from source: include params 13731 1727203841.39506: variable 'bond_port_profile' from source: include params 13731 1727203841.39535: variable 'omit' from source: magic vars 13731 1727203841.39582: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203841.39621: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203841.39782: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203841.39785: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.39788: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.39791: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203841.39793: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.39795: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.39832: Set connection var ansible_pipelining to False 13731 1727203841.39844: Set connection var ansible_shell_type to sh 13731 1727203841.39855: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203841.39865: Set connection var ansible_connection to ssh 13731 1727203841.39877: Set connection var ansible_shell_executable to /bin/sh 13731 1727203841.39888: Set connection var ansible_timeout to 10 13731 1727203841.39918: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.39926: variable 'ansible_connection' from source: unknown 13731 1727203841.39933: variable 'ansible_module_compression' from source: unknown 13731 1727203841.39940: variable 'ansible_shell_type' from source: unknown 13731 1727203841.39948: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.39955: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.39962: variable 'ansible_pipelining' from source: unknown 13731 1727203841.39969: variable 'ansible_timeout' from source: unknown 13731 1727203841.39979: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.40116: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203841.40134: variable 'omit' from source: magic vars 13731 1727203841.40235: starting attempt loop 13731 1727203841.40238: running the handler 13731 1727203841.40248: variable 'lsr_net_profile_exists' from source: set_fact 13731 1727203841.40257: Evaluated conditional (lsr_net_profile_exists): True 13731 1727203841.40266: handler run complete 13731 1727203841.40284: attempt loop complete, returning result 13731 1727203841.40290: _execute() done 13731 1727203841.40295: dumping result to json 13731 1727203841.40301: done dumping result, returning 13731 1727203841.40309: done running TaskExecutor() for managed-node3/TASK: Assert that the profile is present - 'bond0' [028d2410-947f-82dc-c122-0000000004da] 13731 1727203841.40318: sending task result for task 028d2410-947f-82dc-c122-0000000004da ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203841.40497: no more pending results, returning what we have 13731 1727203841.40501: results queue empty 13731 1727203841.40502: checking for any_errors_fatal 13731 1727203841.40510: done checking for any_errors_fatal 13731 1727203841.40511: checking for max_fail_percentage 13731 1727203841.40513: done checking for max_fail_percentage 13731 1727203841.40514: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.40514: done checking to see if all hosts have failed 13731 1727203841.40515: getting the remaining hosts for this loop 13731 1727203841.40517: done getting the remaining hosts for this loop 13731 1727203841.40521: getting the next task for host managed-node3 13731 1727203841.40528: done getting next task for host managed-node3 13731 1727203841.40531: ^ task is: TASK: Assert that the ansible managed comment is present in '{{ profile }}' 13731 1727203841.40536: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.40540: getting variables 13731 1727203841.40541: in VariableManager get_vars() 13731 1727203841.40573: Calling all_inventory to load vars for managed-node3 13731 1727203841.40578: Calling groups_inventory to load vars for managed-node3 13731 1727203841.40581: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.40592: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.40596: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.40599: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.41289: done sending task result for task 028d2410-947f-82dc-c122-0000000004da 13731 1727203841.41292: WORKER PROCESS EXITING 13731 1727203841.42279: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.43781: done with get_vars() 13731 1727203841.43806: done getting variables 13731 1727203841.43869: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203841.43994: variable 'profile' from source: include params 13731 1727203841.43999: variable 'bond_port_profile' from source: include params 13731 1727203841.44058: variable 'bond_port_profile' from source: include params TASK [Assert that the ansible managed comment is present in 'bond0'] *********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:10 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.062) 0:00:19.671 ***** 13731 1727203841.44091: entering _queue_task() for managed-node3/assert 13731 1727203841.44513: worker is 1 (out of 1 available) 13731 1727203841.44524: exiting _queue_task() for managed-node3/assert 13731 1727203841.44535: done queuing things up, now waiting for results queue to drain 13731 1727203841.44536: waiting for pending results... 13731 1727203841.44730: running TaskExecutor() for managed-node3/TASK: Assert that the ansible managed comment is present in 'bond0' 13731 1727203841.44872: in run() - task 028d2410-947f-82dc-c122-0000000004db 13731 1727203841.44901: variable 'ansible_search_path' from source: unknown 13731 1727203841.44910: variable 'ansible_search_path' from source: unknown 13731 1727203841.44952: calling self._execute() 13731 1727203841.45056: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.45069: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.45087: variable 'omit' from source: magic vars 13731 1727203841.45442: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.45461: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.45472: variable 'omit' from source: magic vars 13731 1727203841.45539: variable 'omit' from source: magic vars 13731 1727203841.45641: variable 'profile' from source: include params 13731 1727203841.45651: variable 'bond_port_profile' from source: include params 13731 1727203841.45716: variable 'bond_port_profile' from source: include params 13731 1727203841.45745: variable 'omit' from source: magic vars 13731 1727203841.45793: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203841.45830: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203841.45862: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203841.45888: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.45905: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.45942: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203841.45955: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.45963: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.46066: Set connection var ansible_pipelining to False 13731 1727203841.46082: Set connection var ansible_shell_type to sh 13731 1727203841.46092: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203841.46102: Set connection var ansible_connection to ssh 13731 1727203841.46111: Set connection var ansible_shell_executable to /bin/sh 13731 1727203841.46120: Set connection var ansible_timeout to 10 13731 1727203841.46143: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.46149: variable 'ansible_connection' from source: unknown 13731 1727203841.46155: variable 'ansible_module_compression' from source: unknown 13731 1727203841.46160: variable 'ansible_shell_type' from source: unknown 13731 1727203841.46166: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.46172: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.46281: variable 'ansible_pipelining' from source: unknown 13731 1727203841.46286: variable 'ansible_timeout' from source: unknown 13731 1727203841.46288: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.46337: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203841.46352: variable 'omit' from source: magic vars 13731 1727203841.46363: starting attempt loop 13731 1727203841.46370: running the handler 13731 1727203841.46486: variable 'lsr_net_profile_ansible_managed' from source: set_fact 13731 1727203841.46496: Evaluated conditional (lsr_net_profile_ansible_managed): True 13731 1727203841.46507: handler run complete 13731 1727203841.46529: attempt loop complete, returning result 13731 1727203841.46537: _execute() done 13731 1727203841.46544: dumping result to json 13731 1727203841.46552: done dumping result, returning 13731 1727203841.46564: done running TaskExecutor() for managed-node3/TASK: Assert that the ansible managed comment is present in 'bond0' [028d2410-947f-82dc-c122-0000000004db] 13731 1727203841.46574: sending task result for task 028d2410-947f-82dc-c122-0000000004db 13731 1727203841.46687: done sending task result for task 028d2410-947f-82dc-c122-0000000004db 13731 1727203841.46690: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203841.46743: no more pending results, returning what we have 13731 1727203841.46747: results queue empty 13731 1727203841.46747: checking for any_errors_fatal 13731 1727203841.46754: done checking for any_errors_fatal 13731 1727203841.46755: checking for max_fail_percentage 13731 1727203841.46756: done checking for max_fail_percentage 13731 1727203841.46758: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.46758: done checking to see if all hosts have failed 13731 1727203841.46759: getting the remaining hosts for this loop 13731 1727203841.46760: done getting the remaining hosts for this loop 13731 1727203841.46764: getting the next task for host managed-node3 13731 1727203841.46771: done getting next task for host managed-node3 13731 1727203841.46774: ^ task is: TASK: Assert that the fingerprint comment is present in {{ profile }} 13731 1727203841.46780: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.46784: getting variables 13731 1727203841.46786: in VariableManager get_vars() 13731 1727203841.46816: Calling all_inventory to load vars for managed-node3 13731 1727203841.46818: Calling groups_inventory to load vars for managed-node3 13731 1727203841.46822: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.46832: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.46834: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.46837: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.47631: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.48486: done with get_vars() 13731 1727203841.48510: done getting variables 13731 1727203841.48566: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203841.48684: variable 'profile' from source: include params 13731 1727203841.48687: variable 'bond_port_profile' from source: include params 13731 1727203841.48744: variable 'bond_port_profile' from source: include params TASK [Assert that the fingerprint comment is present in bond0] ***************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:15 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.046) 0:00:19.718 ***** 13731 1727203841.48774: entering _queue_task() for managed-node3/assert 13731 1727203841.49280: worker is 1 (out of 1 available) 13731 1727203841.49290: exiting _queue_task() for managed-node3/assert 13731 1727203841.49300: done queuing things up, now waiting for results queue to drain 13731 1727203841.49301: waiting for pending results... 13731 1727203841.49397: running TaskExecutor() for managed-node3/TASK: Assert that the fingerprint comment is present in bond0 13731 1727203841.49499: in run() - task 028d2410-947f-82dc-c122-0000000004dc 13731 1727203841.49595: variable 'ansible_search_path' from source: unknown 13731 1727203841.49599: variable 'ansible_search_path' from source: unknown 13731 1727203841.49602: calling self._execute() 13731 1727203841.49680: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.49684: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.49686: variable 'omit' from source: magic vars 13731 1727203841.49945: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.49955: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.49964: variable 'omit' from source: magic vars 13731 1727203841.50001: variable 'omit' from source: magic vars 13731 1727203841.50066: variable 'profile' from source: include params 13731 1727203841.50080: variable 'bond_port_profile' from source: include params 13731 1727203841.50122: variable 'bond_port_profile' from source: include params 13731 1727203841.50136: variable 'omit' from source: magic vars 13731 1727203841.50181: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203841.50196: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203841.50212: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203841.50225: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.50234: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.50259: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203841.50264: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.50267: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.50339: Set connection var ansible_pipelining to False 13731 1727203841.50342: Set connection var ansible_shell_type to sh 13731 1727203841.50348: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203841.50353: Set connection var ansible_connection to ssh 13731 1727203841.50358: Set connection var ansible_shell_executable to /bin/sh 13731 1727203841.50365: Set connection var ansible_timeout to 10 13731 1727203841.50380: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.50383: variable 'ansible_connection' from source: unknown 13731 1727203841.50385: variable 'ansible_module_compression' from source: unknown 13731 1727203841.50388: variable 'ansible_shell_type' from source: unknown 13731 1727203841.50390: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.50393: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.50397: variable 'ansible_pipelining' from source: unknown 13731 1727203841.50399: variable 'ansible_timeout' from source: unknown 13731 1727203841.50405: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.50504: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203841.50515: variable 'omit' from source: magic vars 13731 1727203841.50518: starting attempt loop 13731 1727203841.50522: running the handler 13731 1727203841.50596: variable 'lsr_net_profile_fingerprint' from source: set_fact 13731 1727203841.50599: Evaluated conditional (lsr_net_profile_fingerprint): True 13731 1727203841.50605: handler run complete 13731 1727203841.50615: attempt loop complete, returning result 13731 1727203841.50619: _execute() done 13731 1727203841.50622: dumping result to json 13731 1727203841.50624: done dumping result, returning 13731 1727203841.50630: done running TaskExecutor() for managed-node3/TASK: Assert that the fingerprint comment is present in bond0 [028d2410-947f-82dc-c122-0000000004dc] 13731 1727203841.50638: sending task result for task 028d2410-947f-82dc-c122-0000000004dc 13731 1727203841.50714: done sending task result for task 028d2410-947f-82dc-c122-0000000004dc 13731 1727203841.50717: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203841.50781: no more pending results, returning what we have 13731 1727203841.50785: results queue empty 13731 1727203841.50786: checking for any_errors_fatal 13731 1727203841.50794: done checking for any_errors_fatal 13731 1727203841.50794: checking for max_fail_percentage 13731 1727203841.50795: done checking for max_fail_percentage 13731 1727203841.50797: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.50797: done checking to see if all hosts have failed 13731 1727203841.50798: getting the remaining hosts for this loop 13731 1727203841.50800: done getting the remaining hosts for this loop 13731 1727203841.50803: getting the next task for host managed-node3 13731 1727203841.50812: done getting next task for host managed-node3 13731 1727203841.50815: ^ task is: TASK: Include the task 'get_profile_stat.yml' 13731 1727203841.50819: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.50822: getting variables 13731 1727203841.50824: in VariableManager get_vars() 13731 1727203841.50849: Calling all_inventory to load vars for managed-node3 13731 1727203841.50852: Calling groups_inventory to load vars for managed-node3 13731 1727203841.50855: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.50866: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.50868: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.50870: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.55279: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.56616: done with get_vars() 13731 1727203841.56639: done getting variables TASK [Include the task 'get_profile_stat.yml'] ********************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:3 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.079) 0:00:19.798 ***** 13731 1727203841.56731: entering _queue_task() for managed-node3/include_tasks 13731 1727203841.57142: worker is 1 (out of 1 available) 13731 1727203841.57160: exiting _queue_task() for managed-node3/include_tasks 13731 1727203841.57171: done queuing things up, now waiting for results queue to drain 13731 1727203841.57173: waiting for pending results... 13731 1727203841.57372: running TaskExecutor() for managed-node3/TASK: Include the task 'get_profile_stat.yml' 13731 1727203841.57459: in run() - task 028d2410-947f-82dc-c122-0000000004e0 13731 1727203841.57478: variable 'ansible_search_path' from source: unknown 13731 1727203841.57482: variable 'ansible_search_path' from source: unknown 13731 1727203841.57511: calling self._execute() 13731 1727203841.57589: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.57596: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.57603: variable 'omit' from source: magic vars 13731 1727203841.57887: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.57896: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.57902: _execute() done 13731 1727203841.57906: dumping result to json 13731 1727203841.57909: done dumping result, returning 13731 1727203841.57914: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_profile_stat.yml' [028d2410-947f-82dc-c122-0000000004e0] 13731 1727203841.57919: sending task result for task 028d2410-947f-82dc-c122-0000000004e0 13731 1727203841.58011: done sending task result for task 028d2410-947f-82dc-c122-0000000004e0 13731 1727203841.58014: WORKER PROCESS EXITING 13731 1727203841.58068: no more pending results, returning what we have 13731 1727203841.58074: in VariableManager get_vars() 13731 1727203841.58110: Calling all_inventory to load vars for managed-node3 13731 1727203841.58113: Calling groups_inventory to load vars for managed-node3 13731 1727203841.58116: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.58129: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.58131: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.58134: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.58905: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.60207: done with get_vars() 13731 1727203841.60231: variable 'ansible_search_path' from source: unknown 13731 1727203841.60233: variable 'ansible_search_path' from source: unknown 13731 1727203841.60278: we have included files to process 13731 1727203841.60279: generating all_blocks data 13731 1727203841.60280: done generating all_blocks data 13731 1727203841.60284: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203841.60285: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203841.60286: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203841.60902: done processing included file 13731 1727203841.60904: iterating over new_blocks loaded from include file 13731 1727203841.60904: in VariableManager get_vars() 13731 1727203841.60915: done with get_vars() 13731 1727203841.60916: filtering new block on tags 13731 1727203841.60960: done filtering new block on tags 13731 1727203841.60964: in VariableManager get_vars() 13731 1727203841.60977: done with get_vars() 13731 1727203841.60978: filtering new block on tags 13731 1727203841.61013: done filtering new block on tags 13731 1727203841.61014: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml for managed-node3 13731 1727203841.61018: extending task lists for all hosts with included blocks 13731 1727203841.61345: done extending task lists 13731 1727203841.61346: done processing included files 13731 1727203841.61347: results queue empty 13731 1727203841.61347: checking for any_errors_fatal 13731 1727203841.61350: done checking for any_errors_fatal 13731 1727203841.61351: checking for max_fail_percentage 13731 1727203841.61351: done checking for max_fail_percentage 13731 1727203841.61352: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.61353: done checking to see if all hosts have failed 13731 1727203841.61353: getting the remaining hosts for this loop 13731 1727203841.61354: done getting the remaining hosts for this loop 13731 1727203841.61355: getting the next task for host managed-node3 13731 1727203841.61358: done getting next task for host managed-node3 13731 1727203841.61360: ^ task is: TASK: Initialize NM profile exist and ansible_managed comment flag 13731 1727203841.61364: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.61366: getting variables 13731 1727203841.61366: in VariableManager get_vars() 13731 1727203841.61373: Calling all_inventory to load vars for managed-node3 13731 1727203841.61374: Calling groups_inventory to load vars for managed-node3 13731 1727203841.61377: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.61381: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.61383: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.61384: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.61988: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.62823: done with get_vars() 13731 1727203841.62837: done getting variables 13731 1727203841.62867: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Initialize NM profile exist and ansible_managed comment flag] ************ task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:3 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.061) 0:00:19.859 ***** 13731 1727203841.62890: entering _queue_task() for managed-node3/set_fact 13731 1727203841.63132: worker is 1 (out of 1 available) 13731 1727203841.63145: exiting _queue_task() for managed-node3/set_fact 13731 1727203841.63158: done queuing things up, now waiting for results queue to drain 13731 1727203841.63159: waiting for pending results... 13731 1727203841.63339: running TaskExecutor() for managed-node3/TASK: Initialize NM profile exist and ansible_managed comment flag 13731 1727203841.63422: in run() - task 028d2410-947f-82dc-c122-000000000558 13731 1727203841.63435: variable 'ansible_search_path' from source: unknown 13731 1727203841.63439: variable 'ansible_search_path' from source: unknown 13731 1727203841.63471: calling self._execute() 13731 1727203841.63545: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.63548: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.63558: variable 'omit' from source: magic vars 13731 1727203841.63834: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.63844: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.63850: variable 'omit' from source: magic vars 13731 1727203841.63891: variable 'omit' from source: magic vars 13731 1727203841.63913: variable 'omit' from source: magic vars 13731 1727203841.63946: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203841.63974: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203841.63993: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203841.64006: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.64016: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.64040: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203841.64045: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.64048: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.64118: Set connection var ansible_pipelining to False 13731 1727203841.64121: Set connection var ansible_shell_type to sh 13731 1727203841.64127: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203841.64132: Set connection var ansible_connection to ssh 13731 1727203841.64137: Set connection var ansible_shell_executable to /bin/sh 13731 1727203841.64144: Set connection var ansible_timeout to 10 13731 1727203841.64166: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.64169: variable 'ansible_connection' from source: unknown 13731 1727203841.64173: variable 'ansible_module_compression' from source: unknown 13731 1727203841.64177: variable 'ansible_shell_type' from source: unknown 13731 1727203841.64180: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.64183: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.64185: variable 'ansible_pipelining' from source: unknown 13731 1727203841.64187: variable 'ansible_timeout' from source: unknown 13731 1727203841.64190: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.64290: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203841.64298: variable 'omit' from source: magic vars 13731 1727203841.64303: starting attempt loop 13731 1727203841.64306: running the handler 13731 1727203841.64317: handler run complete 13731 1727203841.64325: attempt loop complete, returning result 13731 1727203841.64327: _execute() done 13731 1727203841.64330: dumping result to json 13731 1727203841.64332: done dumping result, returning 13731 1727203841.64338: done running TaskExecutor() for managed-node3/TASK: Initialize NM profile exist and ansible_managed comment flag [028d2410-947f-82dc-c122-000000000558] 13731 1727203841.64344: sending task result for task 028d2410-947f-82dc-c122-000000000558 13731 1727203841.64421: done sending task result for task 028d2410-947f-82dc-c122-000000000558 13731 1727203841.64424: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "lsr_net_profile_ansible_managed": false, "lsr_net_profile_exists": false, "lsr_net_profile_fingerprint": false }, "changed": false } 13731 1727203841.64513: no more pending results, returning what we have 13731 1727203841.64516: results queue empty 13731 1727203841.64517: checking for any_errors_fatal 13731 1727203841.64519: done checking for any_errors_fatal 13731 1727203841.64519: checking for max_fail_percentage 13731 1727203841.64521: done checking for max_fail_percentage 13731 1727203841.64522: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.64522: done checking to see if all hosts have failed 13731 1727203841.64523: getting the remaining hosts for this loop 13731 1727203841.64524: done getting the remaining hosts for this loop 13731 1727203841.64527: getting the next task for host managed-node3 13731 1727203841.64537: done getting next task for host managed-node3 13731 1727203841.64539: ^ task is: TASK: Stat profile file 13731 1727203841.64545: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.64548: getting variables 13731 1727203841.64549: in VariableManager get_vars() 13731 1727203841.64577: Calling all_inventory to load vars for managed-node3 13731 1727203841.64580: Calling groups_inventory to load vars for managed-node3 13731 1727203841.64585: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.64593: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.64595: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.64597: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.65417: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203841.66267: done with get_vars() 13731 1727203841.66284: done getting variables TASK [Stat profile file] ******************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:9 Tuesday 24 September 2024 14:50:41 -0400 (0:00:00.034) 0:00:19.894 ***** 13731 1727203841.66352: entering _queue_task() for managed-node3/stat 13731 1727203841.66571: worker is 1 (out of 1 available) 13731 1727203841.66585: exiting _queue_task() for managed-node3/stat 13731 1727203841.66598: done queuing things up, now waiting for results queue to drain 13731 1727203841.66599: waiting for pending results... 13731 1727203841.66766: running TaskExecutor() for managed-node3/TASK: Stat profile file 13731 1727203841.66836: in run() - task 028d2410-947f-82dc-c122-000000000559 13731 1727203841.66850: variable 'ansible_search_path' from source: unknown 13731 1727203841.66853: variable 'ansible_search_path' from source: unknown 13731 1727203841.66881: calling self._execute() 13731 1727203841.66956: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.66959: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.66969: variable 'omit' from source: magic vars 13731 1727203841.67239: variable 'ansible_distribution_major_version' from source: facts 13731 1727203841.67248: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203841.67255: variable 'omit' from source: magic vars 13731 1727203841.67299: variable 'omit' from source: magic vars 13731 1727203841.67366: variable 'profile' from source: include params 13731 1727203841.67370: variable 'bond_port_profile' from source: include params 13731 1727203841.67414: variable 'bond_port_profile' from source: include params 13731 1727203841.67428: variable 'omit' from source: magic vars 13731 1727203841.67463: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203841.67491: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203841.67506: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203841.67519: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.67529: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203841.67551: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203841.67554: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.67557: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.67627: Set connection var ansible_pipelining to False 13731 1727203841.67631: Set connection var ansible_shell_type to sh 13731 1727203841.67636: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203841.67641: Set connection var ansible_connection to ssh 13731 1727203841.67646: Set connection var ansible_shell_executable to /bin/sh 13731 1727203841.67651: Set connection var ansible_timeout to 10 13731 1727203841.67667: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.67670: variable 'ansible_connection' from source: unknown 13731 1727203841.67673: variable 'ansible_module_compression' from source: unknown 13731 1727203841.67676: variable 'ansible_shell_type' from source: unknown 13731 1727203841.67679: variable 'ansible_shell_executable' from source: unknown 13731 1727203841.67681: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203841.67686: variable 'ansible_pipelining' from source: unknown 13731 1727203841.67688: variable 'ansible_timeout' from source: unknown 13731 1727203841.67692: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203841.67831: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203841.67840: variable 'omit' from source: magic vars 13731 1727203841.67845: starting attempt loop 13731 1727203841.67848: running the handler 13731 1727203841.67863: _low_level_execute_command(): starting 13731 1727203841.67869: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203841.68355: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203841.68389: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.68393: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203841.68395: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.68446: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203841.68449: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203841.68452: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203841.68502: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203841.70166: stdout chunk (state=3): >>>/root <<< 13731 1727203841.70266: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203841.70294: stderr chunk (state=3): >>><<< 13731 1727203841.70297: stdout chunk (state=3): >>><<< 13731 1727203841.70317: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203841.70328: _low_level_execute_command(): starting 13731 1727203841.70333: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471 `" && echo ansible-tmp-1727203841.703169-14949-120513716565471="` echo /root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471 `" ) && sleep 0' 13731 1727203841.70753: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203841.70757: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.70766: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203841.70769: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203841.70771: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.70815: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203841.70820: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203841.70822: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203841.70854: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203841.73170: stdout chunk (state=3): >>>ansible-tmp-1727203841.703169-14949-120513716565471=/root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471 <<< 13731 1727203841.73278: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203841.73300: stderr chunk (state=3): >>><<< 13731 1727203841.73303: stdout chunk (state=3): >>><<< 13731 1727203841.73318: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203841.703169-14949-120513716565471=/root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203841.73350: variable 'ansible_module_compression' from source: unknown 13731 1727203841.73396: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13731 1727203841.73427: variable 'ansible_facts' from source: unknown 13731 1727203841.73490: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/AnsiballZ_stat.py 13731 1727203841.73586: Sending initial data 13731 1727203841.73589: Sent initial data (152 bytes) 13731 1727203841.74016: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203841.74019: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203841.74021: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.74023: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203841.74025: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203841.74027: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.74075: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203841.74085: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203841.74113: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203841.75624: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" <<< 13731 1727203841.75636: stderr chunk (state=3): >>>debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203841.75656: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203841.75697: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp7jn_znpu /root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/AnsiballZ_stat.py <<< 13731 1727203841.75701: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/AnsiballZ_stat.py" <<< 13731 1727203841.75725: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp7jn_znpu" to remote "/root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/AnsiballZ_stat.py" <<< 13731 1727203841.75732: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/AnsiballZ_stat.py" <<< 13731 1727203841.76239: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203841.76282: stderr chunk (state=3): >>><<< 13731 1727203841.76285: stdout chunk (state=3): >>><<< 13731 1727203841.76302: done transferring module to remote 13731 1727203841.76310: _low_level_execute_command(): starting 13731 1727203841.76314: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/ /root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/AnsiballZ_stat.py && sleep 0' 13731 1727203841.76743: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203841.76746: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203841.76748: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.76750: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203841.76756: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.76811: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203841.76815: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203841.76848: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203841.78572: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203841.78594: stderr chunk (state=3): >>><<< 13731 1727203841.78597: stdout chunk (state=3): >>><<< 13731 1727203841.78608: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203841.78611: _low_level_execute_command(): starting 13731 1727203841.78619: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/AnsiballZ_stat.py && sleep 0' 13731 1727203841.79032: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203841.79036: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203841.79038: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203841.79040: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203841.79042: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.79092: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203841.79095: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203841.79137: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203841.94662: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0.0", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13731 1727203841.95834: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203841.95860: stderr chunk (state=3): >>><<< 13731 1727203841.95866: stdout chunk (state=3): >>><<< 13731 1727203841.95883: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0.0", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203841.95908: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/etc/sysconfig/network-scripts/ifcfg-bond0.0', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203841.95914: _low_level_execute_command(): starting 13731 1727203841.95920: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203841.703169-14949-120513716565471/ > /dev/null 2>&1 && sleep 0' 13731 1727203841.96352: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203841.96357: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203841.96393: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.96396: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203841.96398: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203841.96404: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203841.96452: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203841.96456: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203841.96463: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203841.96498: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203841.98313: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203841.98336: stderr chunk (state=3): >>><<< 13731 1727203841.98339: stdout chunk (state=3): >>><<< 13731 1727203841.98351: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203841.98357: handler run complete 13731 1727203841.98375: attempt loop complete, returning result 13731 1727203841.98379: _execute() done 13731 1727203841.98382: dumping result to json 13731 1727203841.98384: done dumping result, returning 13731 1727203841.98395: done running TaskExecutor() for managed-node3/TASK: Stat profile file [028d2410-947f-82dc-c122-000000000559] 13731 1727203841.98397: sending task result for task 028d2410-947f-82dc-c122-000000000559 13731 1727203841.98494: done sending task result for task 028d2410-947f-82dc-c122-000000000559 13731 1727203841.98496: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } 13731 1727203841.98551: no more pending results, returning what we have 13731 1727203841.98555: results queue empty 13731 1727203841.98556: checking for any_errors_fatal 13731 1727203841.98567: done checking for any_errors_fatal 13731 1727203841.98567: checking for max_fail_percentage 13731 1727203841.98569: done checking for max_fail_percentage 13731 1727203841.98570: checking to see if all hosts have failed and the running result is not ok 13731 1727203841.98571: done checking to see if all hosts have failed 13731 1727203841.98572: getting the remaining hosts for this loop 13731 1727203841.98573: done getting the remaining hosts for this loop 13731 1727203841.98578: getting the next task for host managed-node3 13731 1727203841.98586: done getting next task for host managed-node3 13731 1727203841.98588: ^ task is: TASK: Set NM profile exist flag based on the profile files 13731 1727203841.98594: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203841.98598: getting variables 13731 1727203841.98599: in VariableManager get_vars() 13731 1727203841.98630: Calling all_inventory to load vars for managed-node3 13731 1727203841.98633: Calling groups_inventory to load vars for managed-node3 13731 1727203841.98636: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203841.98646: Calling all_plugins_play to load vars for managed-node3 13731 1727203841.98649: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203841.98651: Calling groups_plugins_play to load vars for managed-node3 13731 1727203841.99593: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203842.01121: done with get_vars() 13731 1727203842.01147: done getting variables 13731 1727203842.01212: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag based on the profile files] ******************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:17 Tuesday 24 September 2024 14:50:42 -0400 (0:00:00.348) 0:00:20.243 ***** 13731 1727203842.01249: entering _queue_task() for managed-node3/set_fact 13731 1727203842.01578: worker is 1 (out of 1 available) 13731 1727203842.01590: exiting _queue_task() for managed-node3/set_fact 13731 1727203842.01603: done queuing things up, now waiting for results queue to drain 13731 1727203842.01605: waiting for pending results... 13731 1727203842.01896: running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag based on the profile files 13731 1727203842.02040: in run() - task 028d2410-947f-82dc-c122-00000000055a 13731 1727203842.02045: variable 'ansible_search_path' from source: unknown 13731 1727203842.02049: variable 'ansible_search_path' from source: unknown 13731 1727203842.02149: calling self._execute() 13731 1727203842.02177: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.02184: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.02196: variable 'omit' from source: magic vars 13731 1727203842.02608: variable 'ansible_distribution_major_version' from source: facts 13731 1727203842.02621: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203842.02744: variable 'profile_stat' from source: set_fact 13731 1727203842.02762: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203842.02826: when evaluation is False, skipping this task 13731 1727203842.02830: _execute() done 13731 1727203842.02832: dumping result to json 13731 1727203842.02834: done dumping result, returning 13731 1727203842.02836: done running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag based on the profile files [028d2410-947f-82dc-c122-00000000055a] 13731 1727203842.02838: sending task result for task 028d2410-947f-82dc-c122-00000000055a skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203842.02947: no more pending results, returning what we have 13731 1727203842.02951: results queue empty 13731 1727203842.02952: checking for any_errors_fatal 13731 1727203842.02966: done checking for any_errors_fatal 13731 1727203842.02967: checking for max_fail_percentage 13731 1727203842.02968: done checking for max_fail_percentage 13731 1727203842.02969: checking to see if all hosts have failed and the running result is not ok 13731 1727203842.02970: done checking to see if all hosts have failed 13731 1727203842.02971: getting the remaining hosts for this loop 13731 1727203842.02973: done getting the remaining hosts for this loop 13731 1727203842.02979: getting the next task for host managed-node3 13731 1727203842.02988: done getting next task for host managed-node3 13731 1727203842.02991: ^ task is: TASK: Get NM profile info 13731 1727203842.03000: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203842.03005: getting variables 13731 1727203842.03007: in VariableManager get_vars() 13731 1727203842.03041: Calling all_inventory to load vars for managed-node3 13731 1727203842.03046: Calling groups_inventory to load vars for managed-node3 13731 1727203842.03050: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203842.03066: Calling all_plugins_play to load vars for managed-node3 13731 1727203842.03069: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203842.03073: Calling groups_plugins_play to load vars for managed-node3 13731 1727203842.03698: done sending task result for task 028d2410-947f-82dc-c122-00000000055a 13731 1727203842.03702: WORKER PROCESS EXITING 13731 1727203842.04712: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203842.06311: done with get_vars() 13731 1727203842.06341: done getting variables 13731 1727203842.06406: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Get NM profile info] ***************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:25 Tuesday 24 September 2024 14:50:42 -0400 (0:00:00.051) 0:00:20.295 ***** 13731 1727203842.06444: entering _queue_task() for managed-node3/shell 13731 1727203842.07086: worker is 1 (out of 1 available) 13731 1727203842.07095: exiting _queue_task() for managed-node3/shell 13731 1727203842.07105: done queuing things up, now waiting for results queue to drain 13731 1727203842.07107: waiting for pending results... 13731 1727203842.07247: running TaskExecutor() for managed-node3/TASK: Get NM profile info 13731 1727203842.07330: in run() - task 028d2410-947f-82dc-c122-00000000055b 13731 1727203842.07356: variable 'ansible_search_path' from source: unknown 13731 1727203842.07365: variable 'ansible_search_path' from source: unknown 13731 1727203842.07408: calling self._execute() 13731 1727203842.07533: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.07660: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.07664: variable 'omit' from source: magic vars 13731 1727203842.07979: variable 'ansible_distribution_major_version' from source: facts 13731 1727203842.08007: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203842.08019: variable 'omit' from source: magic vars 13731 1727203842.08091: variable 'omit' from source: magic vars 13731 1727203842.08213: variable 'profile' from source: include params 13731 1727203842.08217: variable 'bond_port_profile' from source: include params 13731 1727203842.08280: variable 'bond_port_profile' from source: include params 13731 1727203842.08322: variable 'omit' from source: magic vars 13731 1727203842.08359: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203842.08418: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203842.08526: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203842.08529: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203842.08533: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203842.08536: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203842.08538: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.08540: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.08685: Set connection var ansible_pipelining to False 13731 1727203842.08698: Set connection var ansible_shell_type to sh 13731 1727203842.08708: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203842.08762: Set connection var ansible_connection to ssh 13731 1727203842.08766: Set connection var ansible_shell_executable to /bin/sh 13731 1727203842.08769: Set connection var ansible_timeout to 10 13731 1727203842.09070: variable 'ansible_shell_executable' from source: unknown 13731 1727203842.09074: variable 'ansible_connection' from source: unknown 13731 1727203842.09078: variable 'ansible_module_compression' from source: unknown 13731 1727203842.09080: variable 'ansible_shell_type' from source: unknown 13731 1727203842.09082: variable 'ansible_shell_executable' from source: unknown 13731 1727203842.09085: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.09087: variable 'ansible_pipelining' from source: unknown 13731 1727203842.09089: variable 'ansible_timeout' from source: unknown 13731 1727203842.09091: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.09229: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203842.09247: variable 'omit' from source: magic vars 13731 1727203842.09257: starting attempt loop 13731 1727203842.09264: running the handler 13731 1727203842.09286: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203842.09315: _low_level_execute_command(): starting 13731 1727203842.09329: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203842.10352: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203842.10421: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203842.10564: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203842.10673: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203842.12347: stdout chunk (state=3): >>>/root <<< 13731 1727203842.12508: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203842.12519: stdout chunk (state=3): >>><<< 13731 1727203842.12564: stderr chunk (state=3): >>><<< 13731 1727203842.12589: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203842.12883: _low_level_execute_command(): starting 13731 1727203842.12887: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154 `" && echo ansible-tmp-1727203842.127838-14973-245582656635154="` echo /root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154 `" ) && sleep 0' 13731 1727203842.13462: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203842.13550: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203842.13579: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203842.13592: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203842.13607: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203842.13652: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203842.16106: stdout chunk (state=3): >>>ansible-tmp-1727203842.127838-14973-245582656635154=/root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154 <<< 13731 1727203842.16249: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203842.16272: stderr chunk (state=3): >>><<< 13731 1727203842.16292: stdout chunk (state=3): >>><<< 13731 1727203842.16316: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203842.127838-14973-245582656635154=/root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203842.16366: variable 'ansible_module_compression' from source: unknown 13731 1727203842.16425: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203842.16479: variable 'ansible_facts' from source: unknown 13731 1727203842.16668: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/AnsiballZ_command.py 13731 1727203842.16797: Sending initial data 13731 1727203842.16800: Sent initial data (155 bytes) 13731 1727203842.17404: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203842.17437: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203842.17451: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203842.17491: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203842.17507: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203842.17598: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203842.17622: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203842.17678: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203842.19215: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203842.19274: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203842.19328: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp6n8wzvrb /root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/AnsiballZ_command.py <<< 13731 1727203842.19331: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/AnsiballZ_command.py" <<< 13731 1727203842.19354: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp6n8wzvrb" to remote "/root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/AnsiballZ_command.py" <<< 13731 1727203842.20031: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203842.20073: stderr chunk (state=3): >>><<< 13731 1727203842.20091: stdout chunk (state=3): >>><<< 13731 1727203842.20127: done transferring module to remote 13731 1727203842.20186: _low_level_execute_command(): starting 13731 1727203842.20189: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/ /root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/AnsiballZ_command.py && sleep 0' 13731 1727203842.20856: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203842.20878: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203842.21004: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203842.21028: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203842.21094: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203842.23101: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203842.23349: stderr chunk (state=3): >>><<< 13731 1727203842.23353: stdout chunk (state=3): >>><<< 13731 1727203842.23355: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203842.23364: _low_level_execute_command(): starting 13731 1727203842.23367: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/AnsiballZ_command.py && sleep 0' 13731 1727203842.24582: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203842.24599: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203842.24673: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203842.24700: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203842.24789: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203842.24810: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203842.24860: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203842.41831: stdout chunk (state=3): >>> {"changed": true, "stdout": "bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "start": "2024-09-24 14:50:42.397477", "end": "2024-09-24 14:50:42.417717", "delta": "0:00:00.020240", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203842.43317: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203842.43417: stderr chunk (state=3): >>><<< 13731 1727203842.43432: stdout chunk (state=3): >>><<< 13731 1727203842.43465: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "start": "2024-09-24 14:50:42.397477", "end": "2024-09-24 14:50:42.417717", "delta": "0:00:00.020240", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203842.43531: done with _execute_module (ansible.legacy.command, {'_raw_params': 'nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203842.43583: _low_level_execute_command(): starting 13731 1727203842.43589: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203842.127838-14973-245582656635154/ > /dev/null 2>&1 && sleep 0' 13731 1727203842.44341: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203842.44359: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203842.44411: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203842.44450: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203842.46293: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203842.46299: stdout chunk (state=3): >>><<< 13731 1727203842.46301: stderr chunk (state=3): >>><<< 13731 1727203842.46322: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203842.46482: handler run complete 13731 1727203842.46487: Evaluated conditional (False): False 13731 1727203842.46491: attempt loop complete, returning result 13731 1727203842.46492: _execute() done 13731 1727203842.46494: dumping result to json 13731 1727203842.46496: done dumping result, returning 13731 1727203842.46498: done running TaskExecutor() for managed-node3/TASK: Get NM profile info [028d2410-947f-82dc-c122-00000000055b] 13731 1727203842.46500: sending task result for task 028d2410-947f-82dc-c122-00000000055b 13731 1727203842.46579: done sending task result for task 028d2410-947f-82dc-c122-00000000055b 13731 1727203842.46583: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.0 | grep /etc", "delta": "0:00:00.020240", "end": "2024-09-24 14:50:42.417717", "rc": 0, "start": "2024-09-24 14:50:42.397477" } STDOUT: bond0.0 /etc/NetworkManager/system-connections/bond0.0.nmconnection 13731 1727203842.46658: no more pending results, returning what we have 13731 1727203842.46666: results queue empty 13731 1727203842.46668: checking for any_errors_fatal 13731 1727203842.46677: done checking for any_errors_fatal 13731 1727203842.46678: checking for max_fail_percentage 13731 1727203842.46679: done checking for max_fail_percentage 13731 1727203842.46680: checking to see if all hosts have failed and the running result is not ok 13731 1727203842.46681: done checking to see if all hosts have failed 13731 1727203842.46682: getting the remaining hosts for this loop 13731 1727203842.46684: done getting the remaining hosts for this loop 13731 1727203842.46687: getting the next task for host managed-node3 13731 1727203842.46696: done getting next task for host managed-node3 13731 1727203842.46699: ^ task is: TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13731 1727203842.46706: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203842.46711: getting variables 13731 1727203842.46712: in VariableManager get_vars() 13731 1727203842.46746: Calling all_inventory to load vars for managed-node3 13731 1727203842.46749: Calling groups_inventory to load vars for managed-node3 13731 1727203842.46752: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203842.46766: Calling all_plugins_play to load vars for managed-node3 13731 1727203842.46769: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203842.46773: Calling groups_plugins_play to load vars for managed-node3 13731 1727203842.49786: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203842.52352: done with get_vars() 13731 1727203842.52492: done getting variables 13731 1727203842.52555: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag and ansible_managed flag true based on the nmcli output] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:35 Tuesday 24 September 2024 14:50:42 -0400 (0:00:00.462) 0:00:20.758 ***** 13731 1727203842.52719: entering _queue_task() for managed-node3/set_fact 13731 1727203842.53201: worker is 1 (out of 1 available) 13731 1727203842.53212: exiting _queue_task() for managed-node3/set_fact 13731 1727203842.53225: done queuing things up, now waiting for results queue to drain 13731 1727203842.53227: waiting for pending results... 13731 1727203842.54092: running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13731 1727203842.54183: in run() - task 028d2410-947f-82dc-c122-00000000055c 13731 1727203842.54187: variable 'ansible_search_path' from source: unknown 13731 1727203842.54190: variable 'ansible_search_path' from source: unknown 13731 1727203842.54192: calling self._execute() 13731 1727203842.54398: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.54411: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.54426: variable 'omit' from source: magic vars 13731 1727203842.55482: variable 'ansible_distribution_major_version' from source: facts 13731 1727203842.55487: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203842.55490: variable 'nm_profile_exists' from source: set_fact 13731 1727203842.55689: Evaluated conditional (nm_profile_exists.rc == 0): True 13731 1727203842.55701: variable 'omit' from source: magic vars 13731 1727203842.55766: variable 'omit' from source: magic vars 13731 1727203842.55805: variable 'omit' from source: magic vars 13731 1727203842.55851: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203842.56281: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203842.56284: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203842.56286: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203842.56288: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203842.56290: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203842.56293: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.56295: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.56535: Set connection var ansible_pipelining to False 13731 1727203842.56548: Set connection var ansible_shell_type to sh 13731 1727203842.56553: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203842.56559: Set connection var ansible_connection to ssh 13731 1727203842.56568: Set connection var ansible_shell_executable to /bin/sh 13731 1727203842.56574: Set connection var ansible_timeout to 10 13731 1727203842.56599: variable 'ansible_shell_executable' from source: unknown 13731 1727203842.56602: variable 'ansible_connection' from source: unknown 13731 1727203842.56604: variable 'ansible_module_compression' from source: unknown 13731 1727203842.56607: variable 'ansible_shell_type' from source: unknown 13731 1727203842.56609: variable 'ansible_shell_executable' from source: unknown 13731 1727203842.56611: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.56613: variable 'ansible_pipelining' from source: unknown 13731 1727203842.56616: variable 'ansible_timeout' from source: unknown 13731 1727203842.56621: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.57081: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203842.57085: variable 'omit' from source: magic vars 13731 1727203842.57087: starting attempt loop 13731 1727203842.57089: running the handler 13731 1727203842.57091: handler run complete 13731 1727203842.57093: attempt loop complete, returning result 13731 1727203842.57095: _execute() done 13731 1727203842.57097: dumping result to json 13731 1727203842.57099: done dumping result, returning 13731 1727203842.57102: done running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output [028d2410-947f-82dc-c122-00000000055c] 13731 1727203842.57104: sending task result for task 028d2410-947f-82dc-c122-00000000055c 13731 1727203842.57182: done sending task result for task 028d2410-947f-82dc-c122-00000000055c 13731 1727203842.57187: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "lsr_net_profile_ansible_managed": true, "lsr_net_profile_exists": true, "lsr_net_profile_fingerprint": true }, "changed": false } 13731 1727203842.57242: no more pending results, returning what we have 13731 1727203842.57246: results queue empty 13731 1727203842.57247: checking for any_errors_fatal 13731 1727203842.57255: done checking for any_errors_fatal 13731 1727203842.57256: checking for max_fail_percentage 13731 1727203842.57258: done checking for max_fail_percentage 13731 1727203842.57259: checking to see if all hosts have failed and the running result is not ok 13731 1727203842.57260: done checking to see if all hosts have failed 13731 1727203842.57260: getting the remaining hosts for this loop 13731 1727203842.57263: done getting the remaining hosts for this loop 13731 1727203842.57266: getting the next task for host managed-node3 13731 1727203842.57281: done getting next task for host managed-node3 13731 1727203842.57284: ^ task is: TASK: Get the ansible_managed comment in ifcfg-{{ profile }} 13731 1727203842.57291: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203842.57295: getting variables 13731 1727203842.57296: in VariableManager get_vars() 13731 1727203842.57327: Calling all_inventory to load vars for managed-node3 13731 1727203842.57330: Calling groups_inventory to load vars for managed-node3 13731 1727203842.57333: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203842.57344: Calling all_plugins_play to load vars for managed-node3 13731 1727203842.57347: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203842.57350: Calling groups_plugins_play to load vars for managed-node3 13731 1727203842.60483: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203842.63751: done with get_vars() 13731 1727203842.63787: done getting variables 13731 1727203842.63848: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203842.64171: variable 'profile' from source: include params 13731 1727203842.64178: variable 'bond_port_profile' from source: include params 13731 1727203842.64239: variable 'bond_port_profile' from source: include params TASK [Get the ansible_managed comment in ifcfg-bond0.0] ************************ task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:49 Tuesday 24 September 2024 14:50:42 -0400 (0:00:00.115) 0:00:20.873 ***** 13731 1727203842.64478: entering _queue_task() for managed-node3/command 13731 1727203842.65033: worker is 1 (out of 1 available) 13731 1727203842.65045: exiting _queue_task() for managed-node3/command 13731 1727203842.65059: done queuing things up, now waiting for results queue to drain 13731 1727203842.65063: waiting for pending results... 13731 1727203842.65544: running TaskExecutor() for managed-node3/TASK: Get the ansible_managed comment in ifcfg-bond0.0 13731 1727203842.65831: in run() - task 028d2410-947f-82dc-c122-00000000055e 13731 1727203842.65853: variable 'ansible_search_path' from source: unknown 13731 1727203842.65887: variable 'ansible_search_path' from source: unknown 13731 1727203842.65997: calling self._execute() 13731 1727203842.66157: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.66281: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.66285: variable 'omit' from source: magic vars 13731 1727203842.66973: variable 'ansible_distribution_major_version' from source: facts 13731 1727203842.66992: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203842.67199: variable 'profile_stat' from source: set_fact 13731 1727203842.67294: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203842.67301: when evaluation is False, skipping this task 13731 1727203842.67308: _execute() done 13731 1727203842.67314: dumping result to json 13731 1727203842.67321: done dumping result, returning 13731 1727203842.67446: done running TaskExecutor() for managed-node3/TASK: Get the ansible_managed comment in ifcfg-bond0.0 [028d2410-947f-82dc-c122-00000000055e] 13731 1727203842.67450: sending task result for task 028d2410-947f-82dc-c122-00000000055e 13731 1727203842.67717: done sending task result for task 028d2410-947f-82dc-c122-00000000055e 13731 1727203842.67722: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203842.67782: no more pending results, returning what we have 13731 1727203842.67787: results queue empty 13731 1727203842.67788: checking for any_errors_fatal 13731 1727203842.67796: done checking for any_errors_fatal 13731 1727203842.67797: checking for max_fail_percentage 13731 1727203842.67799: done checking for max_fail_percentage 13731 1727203842.67800: checking to see if all hosts have failed and the running result is not ok 13731 1727203842.67801: done checking to see if all hosts have failed 13731 1727203842.67802: getting the remaining hosts for this loop 13731 1727203842.67804: done getting the remaining hosts for this loop 13731 1727203842.67808: getting the next task for host managed-node3 13731 1727203842.67819: done getting next task for host managed-node3 13731 1727203842.67822: ^ task is: TASK: Verify the ansible_managed comment in ifcfg-{{ profile }} 13731 1727203842.67830: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203842.67835: getting variables 13731 1727203842.67837: in VariableManager get_vars() 13731 1727203842.67873: Calling all_inventory to load vars for managed-node3 13731 1727203842.68079: Calling groups_inventory to load vars for managed-node3 13731 1727203842.68083: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203842.68093: Calling all_plugins_play to load vars for managed-node3 13731 1727203842.68096: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203842.68098: Calling groups_plugins_play to load vars for managed-node3 13731 1727203842.70626: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203842.74115: done with get_vars() 13731 1727203842.74139: done getting variables 13731 1727203842.74201: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203842.74516: variable 'profile' from source: include params 13731 1727203842.74520: variable 'bond_port_profile' from source: include params 13731 1727203842.74581: variable 'bond_port_profile' from source: include params TASK [Verify the ansible_managed comment in ifcfg-bond0.0] ********************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:56 Tuesday 24 September 2024 14:50:42 -0400 (0:00:00.103) 0:00:20.977 ***** 13731 1727203842.74615: entering _queue_task() for managed-node3/set_fact 13731 1727203842.75272: worker is 1 (out of 1 available) 13731 1727203842.75286: exiting _queue_task() for managed-node3/set_fact 13731 1727203842.75299: done queuing things up, now waiting for results queue to drain 13731 1727203842.75301: waiting for pending results... 13731 1727203842.75756: running TaskExecutor() for managed-node3/TASK: Verify the ansible_managed comment in ifcfg-bond0.0 13731 1727203842.75867: in run() - task 028d2410-947f-82dc-c122-00000000055f 13731 1727203842.75879: variable 'ansible_search_path' from source: unknown 13731 1727203842.76087: variable 'ansible_search_path' from source: unknown 13731 1727203842.76227: calling self._execute() 13731 1727203842.76231: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.76234: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.76236: variable 'omit' from source: magic vars 13731 1727203842.76969: variable 'ansible_distribution_major_version' from source: facts 13731 1727203842.76982: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203842.77308: variable 'profile_stat' from source: set_fact 13731 1727203842.77418: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203842.77421: when evaluation is False, skipping this task 13731 1727203842.77422: _execute() done 13731 1727203842.77424: dumping result to json 13731 1727203842.77426: done dumping result, returning 13731 1727203842.77428: done running TaskExecutor() for managed-node3/TASK: Verify the ansible_managed comment in ifcfg-bond0.0 [028d2410-947f-82dc-c122-00000000055f] 13731 1727203842.77429: sending task result for task 028d2410-947f-82dc-c122-00000000055f 13731 1727203842.77497: done sending task result for task 028d2410-947f-82dc-c122-00000000055f 13731 1727203842.77501: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203842.77553: no more pending results, returning what we have 13731 1727203842.77558: results queue empty 13731 1727203842.77559: checking for any_errors_fatal 13731 1727203842.77570: done checking for any_errors_fatal 13731 1727203842.77571: checking for max_fail_percentage 13731 1727203842.77573: done checking for max_fail_percentage 13731 1727203842.77575: checking to see if all hosts have failed and the running result is not ok 13731 1727203842.77577: done checking to see if all hosts have failed 13731 1727203842.77578: getting the remaining hosts for this loop 13731 1727203842.77580: done getting the remaining hosts for this loop 13731 1727203842.77584: getting the next task for host managed-node3 13731 1727203842.77593: done getting next task for host managed-node3 13731 1727203842.77595: ^ task is: TASK: Get the fingerprint comment in ifcfg-{{ profile }} 13731 1727203842.77602: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203842.77606: getting variables 13731 1727203842.77608: in VariableManager get_vars() 13731 1727203842.77646: Calling all_inventory to load vars for managed-node3 13731 1727203842.77649: Calling groups_inventory to load vars for managed-node3 13731 1727203842.77654: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203842.77670: Calling all_plugins_play to load vars for managed-node3 13731 1727203842.77674: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203842.77784: Calling groups_plugins_play to load vars for managed-node3 13731 1727203842.80513: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203842.83892: done with get_vars() 13731 1727203842.83920: done getting variables 13731 1727203842.84101: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203842.84332: variable 'profile' from source: include params 13731 1727203842.84336: variable 'bond_port_profile' from source: include params 13731 1727203842.84487: variable 'bond_port_profile' from source: include params TASK [Get the fingerprint comment in ifcfg-bond0.0] **************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:62 Tuesday 24 September 2024 14:50:42 -0400 (0:00:00.099) 0:00:21.076 ***** 13731 1727203842.84525: entering _queue_task() for managed-node3/command 13731 1727203842.85309: worker is 1 (out of 1 available) 13731 1727203842.85322: exiting _queue_task() for managed-node3/command 13731 1727203842.85336: done queuing things up, now waiting for results queue to drain 13731 1727203842.85337: waiting for pending results... 13731 1727203842.85941: running TaskExecutor() for managed-node3/TASK: Get the fingerprint comment in ifcfg-bond0.0 13731 1727203842.85946: in run() - task 028d2410-947f-82dc-c122-000000000560 13731 1727203842.86095: variable 'ansible_search_path' from source: unknown 13731 1727203842.86099: variable 'ansible_search_path' from source: unknown 13731 1727203842.86132: calling self._execute() 13731 1727203842.86223: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.86227: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.86237: variable 'omit' from source: magic vars 13731 1727203842.87018: variable 'ansible_distribution_major_version' from source: facts 13731 1727203842.87022: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203842.87321: variable 'profile_stat' from source: set_fact 13731 1727203842.87331: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203842.87334: when evaluation is False, skipping this task 13731 1727203842.87384: _execute() done 13731 1727203842.87387: dumping result to json 13731 1727203842.87389: done dumping result, returning 13731 1727203842.87392: done running TaskExecutor() for managed-node3/TASK: Get the fingerprint comment in ifcfg-bond0.0 [028d2410-947f-82dc-c122-000000000560] 13731 1727203842.87394: sending task result for task 028d2410-947f-82dc-c122-000000000560 13731 1727203842.87681: done sending task result for task 028d2410-947f-82dc-c122-000000000560 13731 1727203842.87684: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203842.87737: no more pending results, returning what we have 13731 1727203842.87741: results queue empty 13731 1727203842.87742: checking for any_errors_fatal 13731 1727203842.87750: done checking for any_errors_fatal 13731 1727203842.87751: checking for max_fail_percentage 13731 1727203842.87753: done checking for max_fail_percentage 13731 1727203842.87753: checking to see if all hosts have failed and the running result is not ok 13731 1727203842.87754: done checking to see if all hosts have failed 13731 1727203842.87755: getting the remaining hosts for this loop 13731 1727203842.87757: done getting the remaining hosts for this loop 13731 1727203842.87760: getting the next task for host managed-node3 13731 1727203842.87770: done getting next task for host managed-node3 13731 1727203842.87773: ^ task is: TASK: Verify the fingerprint comment in ifcfg-{{ profile }} 13731 1727203842.87780: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203842.87783: getting variables 13731 1727203842.87784: in VariableManager get_vars() 13731 1727203842.87816: Calling all_inventory to load vars for managed-node3 13731 1727203842.87819: Calling groups_inventory to load vars for managed-node3 13731 1727203842.87822: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203842.87831: Calling all_plugins_play to load vars for managed-node3 13731 1727203842.87833: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203842.87835: Calling groups_plugins_play to load vars for managed-node3 13731 1727203842.90928: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203842.94070: done with get_vars() 13731 1727203842.94209: done getting variables 13731 1727203842.94355: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203842.94602: variable 'profile' from source: include params 13731 1727203842.94607: variable 'bond_port_profile' from source: include params 13731 1727203842.94788: variable 'bond_port_profile' from source: include params TASK [Verify the fingerprint comment in ifcfg-bond0.0] ************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:69 Tuesday 24 September 2024 14:50:42 -0400 (0:00:00.102) 0:00:21.179 ***** 13731 1727203842.94822: entering _queue_task() for managed-node3/set_fact 13731 1727203842.95742: worker is 1 (out of 1 available) 13731 1727203842.95754: exiting _queue_task() for managed-node3/set_fact 13731 1727203842.95767: done queuing things up, now waiting for results queue to drain 13731 1727203842.95768: waiting for pending results... 13731 1727203842.96160: running TaskExecutor() for managed-node3/TASK: Verify the fingerprint comment in ifcfg-bond0.0 13731 1727203842.96403: in run() - task 028d2410-947f-82dc-c122-000000000561 13731 1727203842.96421: variable 'ansible_search_path' from source: unknown 13731 1727203842.96425: variable 'ansible_search_path' from source: unknown 13731 1727203842.96462: calling self._execute() 13731 1727203842.96617: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203842.96621: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203842.96623: variable 'omit' from source: magic vars 13731 1727203842.96965: variable 'ansible_distribution_major_version' from source: facts 13731 1727203842.96974: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203842.97101: variable 'profile_stat' from source: set_fact 13731 1727203842.97113: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203842.97117: when evaluation is False, skipping this task 13731 1727203842.97120: _execute() done 13731 1727203842.97122: dumping result to json 13731 1727203842.97124: done dumping result, returning 13731 1727203842.97164: done running TaskExecutor() for managed-node3/TASK: Verify the fingerprint comment in ifcfg-bond0.0 [028d2410-947f-82dc-c122-000000000561] 13731 1727203842.97167: sending task result for task 028d2410-947f-82dc-c122-000000000561 13731 1727203842.97230: done sending task result for task 028d2410-947f-82dc-c122-000000000561 13731 1727203842.97233: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203842.97310: no more pending results, returning what we have 13731 1727203842.97315: results queue empty 13731 1727203842.97316: checking for any_errors_fatal 13731 1727203842.97323: done checking for any_errors_fatal 13731 1727203842.97324: checking for max_fail_percentage 13731 1727203842.97325: done checking for max_fail_percentage 13731 1727203842.97326: checking to see if all hosts have failed and the running result is not ok 13731 1727203842.97327: done checking to see if all hosts have failed 13731 1727203842.97328: getting the remaining hosts for this loop 13731 1727203842.97330: done getting the remaining hosts for this loop 13731 1727203842.97333: getting the next task for host managed-node3 13731 1727203842.97342: done getting next task for host managed-node3 13731 1727203842.97345: ^ task is: TASK: Assert that the profile is present - '{{ profile }}' 13731 1727203842.97350: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203842.97354: getting variables 13731 1727203842.97356: in VariableManager get_vars() 13731 1727203842.97391: Calling all_inventory to load vars for managed-node3 13731 1727203842.97395: Calling groups_inventory to load vars for managed-node3 13731 1727203842.97397: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203842.97407: Calling all_plugins_play to load vars for managed-node3 13731 1727203842.97409: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203842.97411: Calling groups_plugins_play to load vars for managed-node3 13731 1727203842.99017: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.00749: done with get_vars() 13731 1727203843.00774: done getting variables 13731 1727203843.00841: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203843.00963: variable 'profile' from source: include params 13731 1727203843.00967: variable 'bond_port_profile' from source: include params 13731 1727203843.01036: variable 'bond_port_profile' from source: include params TASK [Assert that the profile is present - 'bond0.0'] ************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:5 Tuesday 24 September 2024 14:50:43 -0400 (0:00:00.062) 0:00:21.241 ***** 13731 1727203843.01070: entering _queue_task() for managed-node3/assert 13731 1727203843.01410: worker is 1 (out of 1 available) 13731 1727203843.01424: exiting _queue_task() for managed-node3/assert 13731 1727203843.01436: done queuing things up, now waiting for results queue to drain 13731 1727203843.01438: waiting for pending results... 13731 1727203843.01792: running TaskExecutor() for managed-node3/TASK: Assert that the profile is present - 'bond0.0' 13731 1727203843.01914: in run() - task 028d2410-947f-82dc-c122-0000000004e1 13731 1727203843.01918: variable 'ansible_search_path' from source: unknown 13731 1727203843.01921: variable 'ansible_search_path' from source: unknown 13731 1727203843.01927: calling self._execute() 13731 1727203843.02068: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.02071: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.02084: variable 'omit' from source: magic vars 13731 1727203843.02474: variable 'ansible_distribution_major_version' from source: facts 13731 1727203843.02479: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203843.02482: variable 'omit' from source: magic vars 13731 1727203843.02533: variable 'omit' from source: magic vars 13731 1727203843.02781: variable 'profile' from source: include params 13731 1727203843.02785: variable 'bond_port_profile' from source: include params 13731 1727203843.02787: variable 'bond_port_profile' from source: include params 13731 1727203843.02790: variable 'omit' from source: magic vars 13731 1727203843.02792: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203843.02811: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203843.02830: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203843.02847: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.02859: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.02900: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203843.02903: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.02906: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.03283: Set connection var ansible_pipelining to False 13731 1727203843.03286: Set connection var ansible_shell_type to sh 13731 1727203843.03288: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203843.03290: Set connection var ansible_connection to ssh 13731 1727203843.03292: Set connection var ansible_shell_executable to /bin/sh 13731 1727203843.03295: Set connection var ansible_timeout to 10 13731 1727203843.03297: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.03299: variable 'ansible_connection' from source: unknown 13731 1727203843.03301: variable 'ansible_module_compression' from source: unknown 13731 1727203843.03303: variable 'ansible_shell_type' from source: unknown 13731 1727203843.03305: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.03306: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.03308: variable 'ansible_pipelining' from source: unknown 13731 1727203843.03310: variable 'ansible_timeout' from source: unknown 13731 1727203843.03312: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.03315: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203843.03318: variable 'omit' from source: magic vars 13731 1727203843.03321: starting attempt loop 13731 1727203843.03324: running the handler 13731 1727203843.03366: variable 'lsr_net_profile_exists' from source: set_fact 13731 1727203843.03369: Evaluated conditional (lsr_net_profile_exists): True 13731 1727203843.03374: handler run complete 13731 1727203843.03390: attempt loop complete, returning result 13731 1727203843.03393: _execute() done 13731 1727203843.03396: dumping result to json 13731 1727203843.03398: done dumping result, returning 13731 1727203843.03405: done running TaskExecutor() for managed-node3/TASK: Assert that the profile is present - 'bond0.0' [028d2410-947f-82dc-c122-0000000004e1] 13731 1727203843.03410: sending task result for task 028d2410-947f-82dc-c122-0000000004e1 13731 1727203843.03506: done sending task result for task 028d2410-947f-82dc-c122-0000000004e1 13731 1727203843.03509: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203843.03592: no more pending results, returning what we have 13731 1727203843.03596: results queue empty 13731 1727203843.03597: checking for any_errors_fatal 13731 1727203843.03607: done checking for any_errors_fatal 13731 1727203843.03608: checking for max_fail_percentage 13731 1727203843.03610: done checking for max_fail_percentage 13731 1727203843.03611: checking to see if all hosts have failed and the running result is not ok 13731 1727203843.03612: done checking to see if all hosts have failed 13731 1727203843.03613: getting the remaining hosts for this loop 13731 1727203843.03615: done getting the remaining hosts for this loop 13731 1727203843.03618: getting the next task for host managed-node3 13731 1727203843.03627: done getting next task for host managed-node3 13731 1727203843.03629: ^ task is: TASK: Assert that the ansible managed comment is present in '{{ profile }}' 13731 1727203843.03745: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203843.03751: getting variables 13731 1727203843.03752: in VariableManager get_vars() 13731 1727203843.03781: Calling all_inventory to load vars for managed-node3 13731 1727203843.03784: Calling groups_inventory to load vars for managed-node3 13731 1727203843.03788: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203843.03797: Calling all_plugins_play to load vars for managed-node3 13731 1727203843.03800: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203843.03803: Calling groups_plugins_play to load vars for managed-node3 13731 1727203843.05352: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.06916: done with get_vars() 13731 1727203843.06944: done getting variables 13731 1727203843.07005: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203843.07139: variable 'profile' from source: include params 13731 1727203843.07143: variable 'bond_port_profile' from source: include params 13731 1727203843.07202: variable 'bond_port_profile' from source: include params TASK [Assert that the ansible managed comment is present in 'bond0.0'] ********* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:10 Tuesday 24 September 2024 14:50:43 -0400 (0:00:00.061) 0:00:21.303 ***** 13731 1727203843.07239: entering _queue_task() for managed-node3/assert 13731 1727203843.07698: worker is 1 (out of 1 available) 13731 1727203843.07708: exiting _queue_task() for managed-node3/assert 13731 1727203843.07720: done queuing things up, now waiting for results queue to drain 13731 1727203843.07722: waiting for pending results... 13731 1727203843.08079: running TaskExecutor() for managed-node3/TASK: Assert that the ansible managed comment is present in 'bond0.0' 13731 1727203843.08086: in run() - task 028d2410-947f-82dc-c122-0000000004e2 13731 1727203843.08091: variable 'ansible_search_path' from source: unknown 13731 1727203843.08095: variable 'ansible_search_path' from source: unknown 13731 1727203843.08098: calling self._execute() 13731 1727203843.08191: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.08214: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.08219: variable 'omit' from source: magic vars 13731 1727203843.08763: variable 'ansible_distribution_major_version' from source: facts 13731 1727203843.08768: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203843.08771: variable 'omit' from source: magic vars 13731 1727203843.08774: variable 'omit' from source: magic vars 13731 1727203843.08779: variable 'profile' from source: include params 13731 1727203843.08782: variable 'bond_port_profile' from source: include params 13731 1727203843.08845: variable 'bond_port_profile' from source: include params 13731 1727203843.08869: variable 'omit' from source: magic vars 13731 1727203843.08913: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203843.08946: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203843.08966: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203843.08988: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.09000: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.09180: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203843.09183: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.09187: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.09189: Set connection var ansible_pipelining to False 13731 1727203843.09191: Set connection var ansible_shell_type to sh 13731 1727203843.09193: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203843.09195: Set connection var ansible_connection to ssh 13731 1727203843.09197: Set connection var ansible_shell_executable to /bin/sh 13731 1727203843.09200: Set connection var ansible_timeout to 10 13731 1727203843.09202: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.09204: variable 'ansible_connection' from source: unknown 13731 1727203843.09206: variable 'ansible_module_compression' from source: unknown 13731 1727203843.09208: variable 'ansible_shell_type' from source: unknown 13731 1727203843.09210: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.09212: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.09214: variable 'ansible_pipelining' from source: unknown 13731 1727203843.09217: variable 'ansible_timeout' from source: unknown 13731 1727203843.09224: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.09368: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203843.09388: variable 'omit' from source: magic vars 13731 1727203843.09407: starting attempt loop 13731 1727203843.09411: running the handler 13731 1727203843.09549: variable 'lsr_net_profile_ansible_managed' from source: set_fact 13731 1727203843.09553: Evaluated conditional (lsr_net_profile_ansible_managed): True 13731 1727203843.09555: handler run complete 13731 1727203843.09626: attempt loop complete, returning result 13731 1727203843.09629: _execute() done 13731 1727203843.09631: dumping result to json 13731 1727203843.09633: done dumping result, returning 13731 1727203843.09635: done running TaskExecutor() for managed-node3/TASK: Assert that the ansible managed comment is present in 'bond0.0' [028d2410-947f-82dc-c122-0000000004e2] 13731 1727203843.09638: sending task result for task 028d2410-947f-82dc-c122-0000000004e2 13731 1727203843.09911: done sending task result for task 028d2410-947f-82dc-c122-0000000004e2 13731 1727203843.09915: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203843.09960: no more pending results, returning what we have 13731 1727203843.09963: results queue empty 13731 1727203843.09964: checking for any_errors_fatal 13731 1727203843.09969: done checking for any_errors_fatal 13731 1727203843.09969: checking for max_fail_percentage 13731 1727203843.09971: done checking for max_fail_percentage 13731 1727203843.09972: checking to see if all hosts have failed and the running result is not ok 13731 1727203843.09973: done checking to see if all hosts have failed 13731 1727203843.09974: getting the remaining hosts for this loop 13731 1727203843.09977: done getting the remaining hosts for this loop 13731 1727203843.09980: getting the next task for host managed-node3 13731 1727203843.09988: done getting next task for host managed-node3 13731 1727203843.09991: ^ task is: TASK: Assert that the fingerprint comment is present in {{ profile }} 13731 1727203843.09995: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203843.09999: getting variables 13731 1727203843.10000: in VariableManager get_vars() 13731 1727203843.10037: Calling all_inventory to load vars for managed-node3 13731 1727203843.10041: Calling groups_inventory to load vars for managed-node3 13731 1727203843.10045: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203843.10056: Calling all_plugins_play to load vars for managed-node3 13731 1727203843.10059: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203843.10062: Calling groups_plugins_play to load vars for managed-node3 13731 1727203843.11503: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.13187: done with get_vars() 13731 1727203843.13208: done getting variables 13731 1727203843.13266: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203843.13382: variable 'profile' from source: include params 13731 1727203843.13390: variable 'bond_port_profile' from source: include params 13731 1727203843.13449: variable 'bond_port_profile' from source: include params TASK [Assert that the fingerprint comment is present in bond0.0] *************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:15 Tuesday 24 September 2024 14:50:43 -0400 (0:00:00.062) 0:00:21.365 ***** 13731 1727203843.13484: entering _queue_task() for managed-node3/assert 13731 1727203843.13854: worker is 1 (out of 1 available) 13731 1727203843.13866: exiting _queue_task() for managed-node3/assert 13731 1727203843.13936: done queuing things up, now waiting for results queue to drain 13731 1727203843.13939: waiting for pending results... 13731 1727203843.14131: running TaskExecutor() for managed-node3/TASK: Assert that the fingerprint comment is present in bond0.0 13731 1727203843.14272: in run() - task 028d2410-947f-82dc-c122-0000000004e3 13731 1727203843.14296: variable 'ansible_search_path' from source: unknown 13731 1727203843.14304: variable 'ansible_search_path' from source: unknown 13731 1727203843.14343: calling self._execute() 13731 1727203843.14481: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.14484: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.14486: variable 'omit' from source: magic vars 13731 1727203843.14833: variable 'ansible_distribution_major_version' from source: facts 13731 1727203843.14849: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203843.14859: variable 'omit' from source: magic vars 13731 1727203843.14924: variable 'omit' from source: magic vars 13731 1727203843.15280: variable 'profile' from source: include params 13731 1727203843.15284: variable 'bond_port_profile' from source: include params 13731 1727203843.15286: variable 'bond_port_profile' from source: include params 13731 1727203843.15288: variable 'omit' from source: magic vars 13731 1727203843.15290: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203843.15293: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203843.15295: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203843.15297: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.15299: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.15301: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203843.15303: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.15304: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.15398: Set connection var ansible_pipelining to False 13731 1727203843.15408: Set connection var ansible_shell_type to sh 13731 1727203843.15427: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203843.15437: Set connection var ansible_connection to ssh 13731 1727203843.15447: Set connection var ansible_shell_executable to /bin/sh 13731 1727203843.15459: Set connection var ansible_timeout to 10 13731 1727203843.15487: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.15494: variable 'ansible_connection' from source: unknown 13731 1727203843.15501: variable 'ansible_module_compression' from source: unknown 13731 1727203843.15507: variable 'ansible_shell_type' from source: unknown 13731 1727203843.15513: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.15519: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.15534: variable 'ansible_pipelining' from source: unknown 13731 1727203843.15542: variable 'ansible_timeout' from source: unknown 13731 1727203843.15550: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.15700: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203843.15717: variable 'omit' from source: magic vars 13731 1727203843.15727: starting attempt loop 13731 1727203843.15735: running the handler 13731 1727203843.15854: variable 'lsr_net_profile_fingerprint' from source: set_fact 13731 1727203843.15868: Evaluated conditional (lsr_net_profile_fingerprint): True 13731 1727203843.15882: handler run complete 13731 1727203843.15902: attempt loop complete, returning result 13731 1727203843.15967: _execute() done 13731 1727203843.15970: dumping result to json 13731 1727203843.15972: done dumping result, returning 13731 1727203843.15976: done running TaskExecutor() for managed-node3/TASK: Assert that the fingerprint comment is present in bond0.0 [028d2410-947f-82dc-c122-0000000004e3] 13731 1727203843.15979: sending task result for task 028d2410-947f-82dc-c122-0000000004e3 13731 1727203843.16038: done sending task result for task 028d2410-947f-82dc-c122-0000000004e3 13731 1727203843.16040: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203843.16118: no more pending results, returning what we have 13731 1727203843.16122: results queue empty 13731 1727203843.16123: checking for any_errors_fatal 13731 1727203843.16131: done checking for any_errors_fatal 13731 1727203843.16132: checking for max_fail_percentage 13731 1727203843.16134: done checking for max_fail_percentage 13731 1727203843.16135: checking to see if all hosts have failed and the running result is not ok 13731 1727203843.16136: done checking to see if all hosts have failed 13731 1727203843.16136: getting the remaining hosts for this loop 13731 1727203843.16138: done getting the remaining hosts for this loop 13731 1727203843.16142: getting the next task for host managed-node3 13731 1727203843.16155: done getting next task for host managed-node3 13731 1727203843.16160: ^ task is: TASK: Include the task 'get_profile_stat.yml' 13731 1727203843.16164: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203843.16169: getting variables 13731 1727203843.16171: in VariableManager get_vars() 13731 1727203843.16206: Calling all_inventory to load vars for managed-node3 13731 1727203843.16209: Calling groups_inventory to load vars for managed-node3 13731 1727203843.16213: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203843.16223: Calling all_plugins_play to load vars for managed-node3 13731 1727203843.16226: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203843.16229: Calling groups_plugins_play to load vars for managed-node3 13731 1727203843.17826: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.19435: done with get_vars() 13731 1727203843.19464: done getting variables TASK [Include the task 'get_profile_stat.yml'] ********************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:3 Tuesday 24 September 2024 14:50:43 -0400 (0:00:00.060) 0:00:21.426 ***** 13731 1727203843.19574: entering _queue_task() for managed-node3/include_tasks 13731 1727203843.19926: worker is 1 (out of 1 available) 13731 1727203843.19939: exiting _queue_task() for managed-node3/include_tasks 13731 1727203843.19952: done queuing things up, now waiting for results queue to drain 13731 1727203843.19953: waiting for pending results... 13731 1727203843.20306: running TaskExecutor() for managed-node3/TASK: Include the task 'get_profile_stat.yml' 13731 1727203843.20375: in run() - task 028d2410-947f-82dc-c122-0000000004e7 13731 1727203843.20403: variable 'ansible_search_path' from source: unknown 13731 1727203843.20512: variable 'ansible_search_path' from source: unknown 13731 1727203843.20515: calling self._execute() 13731 1727203843.20549: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.20560: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.20577: variable 'omit' from source: magic vars 13731 1727203843.20954: variable 'ansible_distribution_major_version' from source: facts 13731 1727203843.20971: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203843.20985: _execute() done 13731 1727203843.20993: dumping result to json 13731 1727203843.21001: done dumping result, returning 13731 1727203843.21010: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_profile_stat.yml' [028d2410-947f-82dc-c122-0000000004e7] 13731 1727203843.21021: sending task result for task 028d2410-947f-82dc-c122-0000000004e7 13731 1727203843.21186: no more pending results, returning what we have 13731 1727203843.21192: in VariableManager get_vars() 13731 1727203843.21231: Calling all_inventory to load vars for managed-node3 13731 1727203843.21234: Calling groups_inventory to load vars for managed-node3 13731 1727203843.21238: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203843.21252: Calling all_plugins_play to load vars for managed-node3 13731 1727203843.21255: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203843.21257: Calling groups_plugins_play to load vars for managed-node3 13731 1727203843.21991: done sending task result for task 028d2410-947f-82dc-c122-0000000004e7 13731 1727203843.21994: WORKER PROCESS EXITING 13731 1727203843.23055: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.24631: done with get_vars() 13731 1727203843.24653: variable 'ansible_search_path' from source: unknown 13731 1727203843.24655: variable 'ansible_search_path' from source: unknown 13731 1727203843.24695: we have included files to process 13731 1727203843.24697: generating all_blocks data 13731 1727203843.24699: done generating all_blocks data 13731 1727203843.24703: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203843.24704: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203843.24706: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml 13731 1727203843.25645: done processing included file 13731 1727203843.25647: iterating over new_blocks loaded from include file 13731 1727203843.25648: in VariableManager get_vars() 13731 1727203843.25664: done with get_vars() 13731 1727203843.25665: filtering new block on tags 13731 1727203843.25743: done filtering new block on tags 13731 1727203843.25746: in VariableManager get_vars() 13731 1727203843.25762: done with get_vars() 13731 1727203843.25763: filtering new block on tags 13731 1727203843.25824: done filtering new block on tags 13731 1727203843.25828: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml for managed-node3 13731 1727203843.25833: extending task lists for all hosts with included blocks 13731 1727203843.26241: done extending task lists 13731 1727203843.26242: done processing included files 13731 1727203843.26248: results queue empty 13731 1727203843.26249: checking for any_errors_fatal 13731 1727203843.26251: done checking for any_errors_fatal 13731 1727203843.26252: checking for max_fail_percentage 13731 1727203843.26254: done checking for max_fail_percentage 13731 1727203843.26255: checking to see if all hosts have failed and the running result is not ok 13731 1727203843.26256: done checking to see if all hosts have failed 13731 1727203843.26256: getting the remaining hosts for this loop 13731 1727203843.26257: done getting the remaining hosts for this loop 13731 1727203843.26260: getting the next task for host managed-node3 13731 1727203843.26265: done getting next task for host managed-node3 13731 1727203843.26267: ^ task is: TASK: Initialize NM profile exist and ansible_managed comment flag 13731 1727203843.26270: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203843.26272: getting variables 13731 1727203843.26273: in VariableManager get_vars() 13731 1727203843.26283: Calling all_inventory to load vars for managed-node3 13731 1727203843.26286: Calling groups_inventory to load vars for managed-node3 13731 1727203843.26288: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203843.26293: Calling all_plugins_play to load vars for managed-node3 13731 1727203843.26296: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203843.26298: Calling groups_plugins_play to load vars for managed-node3 13731 1727203843.27458: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.29013: done with get_vars() 13731 1727203843.29036: done getting variables 13731 1727203843.29088: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Initialize NM profile exist and ansible_managed comment flag] ************ task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:3 Tuesday 24 September 2024 14:50:43 -0400 (0:00:00.095) 0:00:21.522 ***** 13731 1727203843.29120: entering _queue_task() for managed-node3/set_fact 13731 1727203843.29480: worker is 1 (out of 1 available) 13731 1727203843.29678: exiting _queue_task() for managed-node3/set_fact 13731 1727203843.29690: done queuing things up, now waiting for results queue to drain 13731 1727203843.29691: waiting for pending results... 13731 1727203843.29791: running TaskExecutor() for managed-node3/TASK: Initialize NM profile exist and ansible_managed comment flag 13731 1727203843.29936: in run() - task 028d2410-947f-82dc-c122-0000000005b4 13731 1727203843.29955: variable 'ansible_search_path' from source: unknown 13731 1727203843.29963: variable 'ansible_search_path' from source: unknown 13731 1727203843.30004: calling self._execute() 13731 1727203843.30107: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.30120: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.30140: variable 'omit' from source: magic vars 13731 1727203843.30524: variable 'ansible_distribution_major_version' from source: facts 13731 1727203843.30541: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203843.30552: variable 'omit' from source: magic vars 13731 1727203843.30624: variable 'omit' from source: magic vars 13731 1727203843.30664: variable 'omit' from source: magic vars 13731 1727203843.30716: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203843.30756: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203843.30790: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203843.30813: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.30829: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.30890: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203843.30898: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.30901: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.30987: Set connection var ansible_pipelining to False 13731 1727203843.31006: Set connection var ansible_shell_type to sh 13731 1727203843.31080: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203843.31083: Set connection var ansible_connection to ssh 13731 1727203843.31086: Set connection var ansible_shell_executable to /bin/sh 13731 1727203843.31088: Set connection var ansible_timeout to 10 13731 1727203843.31090: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.31092: variable 'ansible_connection' from source: unknown 13731 1727203843.31095: variable 'ansible_module_compression' from source: unknown 13731 1727203843.31097: variable 'ansible_shell_type' from source: unknown 13731 1727203843.31099: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.31103: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.31105: variable 'ansible_pipelining' from source: unknown 13731 1727203843.31106: variable 'ansible_timeout' from source: unknown 13731 1727203843.31113: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.31257: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203843.31273: variable 'omit' from source: magic vars 13731 1727203843.31286: starting attempt loop 13731 1727203843.31292: running the handler 13731 1727203843.31309: handler run complete 13731 1727203843.31323: attempt loop complete, returning result 13731 1727203843.31337: _execute() done 13731 1727203843.31443: dumping result to json 13731 1727203843.31446: done dumping result, returning 13731 1727203843.31448: done running TaskExecutor() for managed-node3/TASK: Initialize NM profile exist and ansible_managed comment flag [028d2410-947f-82dc-c122-0000000005b4] 13731 1727203843.31451: sending task result for task 028d2410-947f-82dc-c122-0000000005b4 13731 1727203843.31519: done sending task result for task 028d2410-947f-82dc-c122-0000000005b4 13731 1727203843.31522: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "lsr_net_profile_ansible_managed": false, "lsr_net_profile_exists": false, "lsr_net_profile_fingerprint": false }, "changed": false } 13731 1727203843.31599: no more pending results, returning what we have 13731 1727203843.31603: results queue empty 13731 1727203843.31604: checking for any_errors_fatal 13731 1727203843.31606: done checking for any_errors_fatal 13731 1727203843.31607: checking for max_fail_percentage 13731 1727203843.31608: done checking for max_fail_percentage 13731 1727203843.31609: checking to see if all hosts have failed and the running result is not ok 13731 1727203843.31610: done checking to see if all hosts have failed 13731 1727203843.31611: getting the remaining hosts for this loop 13731 1727203843.31614: done getting the remaining hosts for this loop 13731 1727203843.31617: getting the next task for host managed-node3 13731 1727203843.31628: done getting next task for host managed-node3 13731 1727203843.31631: ^ task is: TASK: Stat profile file 13731 1727203843.31637: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203843.31641: getting variables 13731 1727203843.31642: in VariableManager get_vars() 13731 1727203843.31677: Calling all_inventory to load vars for managed-node3 13731 1727203843.31680: Calling groups_inventory to load vars for managed-node3 13731 1727203843.31684: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203843.31695: Calling all_plugins_play to load vars for managed-node3 13731 1727203843.31698: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203843.31701: Calling groups_plugins_play to load vars for managed-node3 13731 1727203843.33329: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.34904: done with get_vars() 13731 1727203843.34930: done getting variables TASK [Stat profile file] ******************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:9 Tuesday 24 September 2024 14:50:43 -0400 (0:00:00.059) 0:00:21.581 ***** 13731 1727203843.35033: entering _queue_task() for managed-node3/stat 13731 1727203843.35446: worker is 1 (out of 1 available) 13731 1727203843.35469: exiting _queue_task() for managed-node3/stat 13731 1727203843.35489: done queuing things up, now waiting for results queue to drain 13731 1727203843.35491: waiting for pending results... 13731 1727203843.35821: running TaskExecutor() for managed-node3/TASK: Stat profile file 13731 1727203843.35916: in run() - task 028d2410-947f-82dc-c122-0000000005b5 13731 1727203843.35937: variable 'ansible_search_path' from source: unknown 13731 1727203843.35946: variable 'ansible_search_path' from source: unknown 13731 1727203843.35987: calling self._execute() 13731 1727203843.36085: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.36097: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.36118: variable 'omit' from source: magic vars 13731 1727203843.36488: variable 'ansible_distribution_major_version' from source: facts 13731 1727203843.36505: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203843.36515: variable 'omit' from source: magic vars 13731 1727203843.36585: variable 'omit' from source: magic vars 13731 1727203843.36759: variable 'profile' from source: include params 13731 1727203843.36762: variable 'bond_port_profile' from source: include params 13731 1727203843.36764: variable 'bond_port_profile' from source: include params 13731 1727203843.36786: variable 'omit' from source: magic vars 13731 1727203843.36830: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203843.36878: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203843.36903: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203843.36924: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.36940: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.36982: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203843.36990: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.36998: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.37108: Set connection var ansible_pipelining to False 13731 1727203843.37118: Set connection var ansible_shell_type to sh 13731 1727203843.37127: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203843.37192: Set connection var ansible_connection to ssh 13731 1727203843.37194: Set connection var ansible_shell_executable to /bin/sh 13731 1727203843.37197: Set connection var ansible_timeout to 10 13731 1727203843.37199: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.37201: variable 'ansible_connection' from source: unknown 13731 1727203843.37204: variable 'ansible_module_compression' from source: unknown 13731 1727203843.37206: variable 'ansible_shell_type' from source: unknown 13731 1727203843.37208: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.37210: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.37211: variable 'ansible_pipelining' from source: unknown 13731 1727203843.37225: variable 'ansible_timeout' from source: unknown 13731 1727203843.37264: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.37523: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203843.37539: variable 'omit' from source: magic vars 13731 1727203843.37548: starting attempt loop 13731 1727203843.37554: running the handler 13731 1727203843.37625: _low_level_execute_command(): starting 13731 1727203843.37628: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203843.38316: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203843.38401: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203843.38445: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203843.38827: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203843.38844: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203843.40682: stdout chunk (state=3): >>>/root <<< 13731 1727203843.40686: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203843.40688: stdout chunk (state=3): >>><<< 13731 1727203843.40690: stderr chunk (state=3): >>><<< 13731 1727203843.40693: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203843.40695: _low_level_execute_command(): starting 13731 1727203843.40698: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403 `" && echo ansible-tmp-1727203843.4062548-15084-129921676736403="` echo /root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403 `" ) && sleep 0' 13731 1727203843.42091: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203843.42157: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203843.42177: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203843.42187: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203843.42195: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203843.42204: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203843.42214: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203843.42226: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203843.42234: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203843.42289: stderr chunk (state=3): >>>debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203843.42425: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203843.42461: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203843.42671: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203843.44500: stdout chunk (state=3): >>>ansible-tmp-1727203843.4062548-15084-129921676736403=/root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403 <<< 13731 1727203843.44542: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203843.44593: stderr chunk (state=3): >>><<< 13731 1727203843.44597: stdout chunk (state=3): >>><<< 13731 1727203843.44668: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203843.4062548-15084-129921676736403=/root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203843.44744: variable 'ansible_module_compression' from source: unknown 13731 1727203843.45182: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13731 1727203843.45185: variable 'ansible_facts' from source: unknown 13731 1727203843.45216: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/AnsiballZ_stat.py 13731 1727203843.45441: Sending initial data 13731 1727203843.45445: Sent initial data (153 bytes) 13731 1727203843.46926: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203843.46935: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203843.46949: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203843.46955: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203843.46968: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203843.46986: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203843.47246: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203843.47250: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203843.47252: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203843.47279: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203843.47336: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203843.48878: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203843.49043: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/AnsiballZ_stat.py" <<< 13731 1727203843.49052: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp4ifh6urm /root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/AnsiballZ_stat.py <<< 13731 1727203843.49086: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp4ifh6urm" to remote "/root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/AnsiballZ_stat.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/AnsiballZ_stat.py" <<< 13731 1727203843.50881: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203843.50884: stdout chunk (state=3): >>><<< 13731 1727203843.50886: stderr chunk (state=3): >>><<< 13731 1727203843.50904: done transferring module to remote 13731 1727203843.50914: _low_level_execute_command(): starting 13731 1727203843.50920: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/ /root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/AnsiballZ_stat.py && sleep 0' 13731 1727203843.52393: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203843.52467: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203843.52471: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203843.52473: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203843.52785: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203843.54417: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203843.54448: stderr chunk (state=3): >>><<< 13731 1727203843.54612: stdout chunk (state=3): >>><<< 13731 1727203843.54615: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203843.54618: _low_level_execute_command(): starting 13731 1727203843.54620: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/AnsiballZ_stat.py && sleep 0' 13731 1727203843.55809: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203843.56002: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203843.56258: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203843.70984: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0.1", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13731 1727203843.72446: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203843.72451: stdout chunk (state=3): >>><<< 13731 1727203843.72456: stderr chunk (state=3): >>><<< 13731 1727203843.72486: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/etc/sysconfig/network-scripts/ifcfg-bond0.1", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203843.72512: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/etc/sysconfig/network-scripts/ifcfg-bond0.1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203843.72521: _low_level_execute_command(): starting 13731 1727203843.72525: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203843.4062548-15084-129921676736403/ > /dev/null 2>&1 && sleep 0' 13731 1727203843.73852: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203843.73885: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203843.73896: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203843.73910: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203843.73923: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203843.74076: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203843.74237: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203843.74247: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203843.76047: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203843.76282: stderr chunk (state=3): >>><<< 13731 1727203843.76285: stdout chunk (state=3): >>><<< 13731 1727203843.76287: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203843.76289: handler run complete 13731 1727203843.76291: attempt loop complete, returning result 13731 1727203843.76292: _execute() done 13731 1727203843.76294: dumping result to json 13731 1727203843.76295: done dumping result, returning 13731 1727203843.76297: done running TaskExecutor() for managed-node3/TASK: Stat profile file [028d2410-947f-82dc-c122-0000000005b5] 13731 1727203843.76298: sending task result for task 028d2410-947f-82dc-c122-0000000005b5 13731 1727203843.76361: done sending task result for task 028d2410-947f-82dc-c122-0000000005b5 13731 1727203843.76364: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } 13731 1727203843.76427: no more pending results, returning what we have 13731 1727203843.76431: results queue empty 13731 1727203843.76432: checking for any_errors_fatal 13731 1727203843.76438: done checking for any_errors_fatal 13731 1727203843.76439: checking for max_fail_percentage 13731 1727203843.76441: done checking for max_fail_percentage 13731 1727203843.76441: checking to see if all hosts have failed and the running result is not ok 13731 1727203843.76442: done checking to see if all hosts have failed 13731 1727203843.76443: getting the remaining hosts for this loop 13731 1727203843.76445: done getting the remaining hosts for this loop 13731 1727203843.76448: getting the next task for host managed-node3 13731 1727203843.76455: done getting next task for host managed-node3 13731 1727203843.76457: ^ task is: TASK: Set NM profile exist flag based on the profile files 13731 1727203843.76462: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203843.76466: getting variables 13731 1727203843.76467: in VariableManager get_vars() 13731 1727203843.76504: Calling all_inventory to load vars for managed-node3 13731 1727203843.76506: Calling groups_inventory to load vars for managed-node3 13731 1727203843.76509: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203843.76519: Calling all_plugins_play to load vars for managed-node3 13731 1727203843.76521: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203843.76523: Calling groups_plugins_play to load vars for managed-node3 13731 1727203843.79398: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.81299: done with get_vars() 13731 1727203843.81321: done getting variables 13731 1727203843.81395: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag based on the profile files] ******************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:17 Tuesday 24 September 2024 14:50:43 -0400 (0:00:00.463) 0:00:22.045 ***** 13731 1727203843.81430: entering _queue_task() for managed-node3/set_fact 13731 1727203843.82536: worker is 1 (out of 1 available) 13731 1727203843.82550: exiting _queue_task() for managed-node3/set_fact 13731 1727203843.82571: done queuing things up, now waiting for results queue to drain 13731 1727203843.82574: waiting for pending results... 13731 1727203843.82822: running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag based on the profile files 13731 1727203843.82936: in run() - task 028d2410-947f-82dc-c122-0000000005b6 13731 1727203843.82943: variable 'ansible_search_path' from source: unknown 13731 1727203843.82946: variable 'ansible_search_path' from source: unknown 13731 1727203843.82949: calling self._execute() 13731 1727203843.83157: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.83163: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.83166: variable 'omit' from source: magic vars 13731 1727203843.83584: variable 'ansible_distribution_major_version' from source: facts 13731 1727203843.83588: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203843.83650: variable 'profile_stat' from source: set_fact 13731 1727203843.83669: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203843.83673: when evaluation is False, skipping this task 13731 1727203843.83679: _execute() done 13731 1727203843.83682: dumping result to json 13731 1727203843.83701: done dumping result, returning 13731 1727203843.83704: done running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag based on the profile files [028d2410-947f-82dc-c122-0000000005b6] 13731 1727203843.83706: sending task result for task 028d2410-947f-82dc-c122-0000000005b6 13731 1727203843.84105: done sending task result for task 028d2410-947f-82dc-c122-0000000005b6 13731 1727203843.84108: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203843.84225: no more pending results, returning what we have 13731 1727203843.84229: results queue empty 13731 1727203843.84230: checking for any_errors_fatal 13731 1727203843.84236: done checking for any_errors_fatal 13731 1727203843.84237: checking for max_fail_percentage 13731 1727203843.84241: done checking for max_fail_percentage 13731 1727203843.84242: checking to see if all hosts have failed and the running result is not ok 13731 1727203843.84242: done checking to see if all hosts have failed 13731 1727203843.84243: getting the remaining hosts for this loop 13731 1727203843.84244: done getting the remaining hosts for this loop 13731 1727203843.84249: getting the next task for host managed-node3 13731 1727203843.84257: done getting next task for host managed-node3 13731 1727203843.84263: ^ task is: TASK: Get NM profile info 13731 1727203843.84268: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203843.84272: getting variables 13731 1727203843.84274: in VariableManager get_vars() 13731 1727203843.84307: Calling all_inventory to load vars for managed-node3 13731 1727203843.84310: Calling groups_inventory to load vars for managed-node3 13731 1727203843.84315: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203843.84325: Calling all_plugins_play to load vars for managed-node3 13731 1727203843.84329: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203843.84332: Calling groups_plugins_play to load vars for managed-node3 13731 1727203843.93003: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203843.95046: done with get_vars() 13731 1727203843.95110: done getting variables 13731 1727203843.95194: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Get NM profile info] ***************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:25 Tuesday 24 September 2024 14:50:43 -0400 (0:00:00.137) 0:00:22.183 ***** 13731 1727203843.95230: entering _queue_task() for managed-node3/shell 13731 1727203843.95769: worker is 1 (out of 1 available) 13731 1727203843.95782: exiting _queue_task() for managed-node3/shell 13731 1727203843.95805: done queuing things up, now waiting for results queue to drain 13731 1727203843.95808: waiting for pending results... 13731 1727203843.96162: running TaskExecutor() for managed-node3/TASK: Get NM profile info 13731 1727203843.96301: in run() - task 028d2410-947f-82dc-c122-0000000005b7 13731 1727203843.96348: variable 'ansible_search_path' from source: unknown 13731 1727203843.96352: variable 'ansible_search_path' from source: unknown 13731 1727203843.96458: calling self._execute() 13731 1727203843.96502: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.96505: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.96509: variable 'omit' from source: magic vars 13731 1727203843.97007: variable 'ansible_distribution_major_version' from source: facts 13731 1727203843.97153: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203843.97158: variable 'omit' from source: magic vars 13731 1727203843.97161: variable 'omit' from source: magic vars 13731 1727203843.97633: variable 'profile' from source: include params 13731 1727203843.97637: variable 'bond_port_profile' from source: include params 13731 1727203843.97640: variable 'bond_port_profile' from source: include params 13731 1727203843.97643: variable 'omit' from source: magic vars 13731 1727203843.97646: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203843.97649: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203843.97651: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203843.97654: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.97656: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203843.97659: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203843.97662: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.97664: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.97744: Set connection var ansible_pipelining to False 13731 1727203843.97750: Set connection var ansible_shell_type to sh 13731 1727203843.97753: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203843.97755: Set connection var ansible_connection to ssh 13731 1727203843.97759: Set connection var ansible_shell_executable to /bin/sh 13731 1727203843.97782: Set connection var ansible_timeout to 10 13731 1727203843.97808: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.97811: variable 'ansible_connection' from source: unknown 13731 1727203843.97825: variable 'ansible_module_compression' from source: unknown 13731 1727203843.97828: variable 'ansible_shell_type' from source: unknown 13731 1727203843.97831: variable 'ansible_shell_executable' from source: unknown 13731 1727203843.97833: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203843.97835: variable 'ansible_pipelining' from source: unknown 13731 1727203843.97838: variable 'ansible_timeout' from source: unknown 13731 1727203843.97840: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203843.98022: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203843.98029: variable 'omit' from source: magic vars 13731 1727203843.98035: starting attempt loop 13731 1727203843.98038: running the handler 13731 1727203843.98056: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203843.98072: _low_level_execute_command(): starting 13731 1727203843.98082: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203843.99107: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203843.99111: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203843.99114: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203843.99191: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203843.99215: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203843.99230: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203843.99321: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203843.99466: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203844.01009: stdout chunk (state=3): >>>/root <<< 13731 1727203844.01189: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203844.01193: stdout chunk (state=3): >>><<< 13731 1727203844.01195: stderr chunk (state=3): >>><<< 13731 1727203844.01302: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203844.01361: _low_level_execute_command(): starting 13731 1727203844.01528: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535 `" && echo ansible-tmp-1727203844.0126579-15109-201459069444535="` echo /root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535 `" ) && sleep 0' 13731 1727203844.02265: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203844.02269: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203844.02271: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203844.02274: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203844.02281: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203844.02288: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203844.02299: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.02323: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203844.02354: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203844.02360: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203844.02363: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203844.02365: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203844.02367: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203844.02376: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203844.02383: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203844.02393: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.02469: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203844.02492: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203844.02498: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203844.02658: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203844.04794: stdout chunk (state=3): >>>ansible-tmp-1727203844.0126579-15109-201459069444535=/root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535 <<< 13731 1727203844.04797: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203844.04800: stdout chunk (state=3): >>><<< 13731 1727203844.04802: stderr chunk (state=3): >>><<< 13731 1727203844.04804: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203844.0126579-15109-201459069444535=/root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203844.04806: variable 'ansible_module_compression' from source: unknown 13731 1727203844.04893: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203844.05018: variable 'ansible_facts' from source: unknown 13731 1727203844.05399: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/AnsiballZ_command.py 13731 1727203844.05821: Sending initial data 13731 1727203844.05849: Sent initial data (156 bytes) 13731 1727203844.07290: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203844.07294: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203844.07297: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203844.07300: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.07456: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203844.07492: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203844.07555: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203844.09208: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203844.09251: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203844.09311: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpz1xo5z8d /root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/AnsiballZ_command.py <<< 13731 1727203844.09331: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/AnsiballZ_command.py" <<< 13731 1727203844.09390: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory <<< 13731 1727203844.09431: stderr chunk (state=3): >>>debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpz1xo5z8d" to remote "/root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/AnsiballZ_command.py" <<< 13731 1727203844.10370: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203844.10512: stderr chunk (state=3): >>><<< 13731 1727203844.10516: stdout chunk (state=3): >>><<< 13731 1727203844.10827: done transferring module to remote 13731 1727203844.10840: _low_level_execute_command(): starting 13731 1727203844.10844: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/ /root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/AnsiballZ_command.py && sleep 0' 13731 1727203844.11514: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203844.11521: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203844.11543: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203844.11546: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203844.11557: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203844.11563: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203844.11594: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.11605: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.11731: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203844.11742: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203844.11778: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203844.13470: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203844.13492: stderr chunk (state=3): >>><<< 13731 1727203844.13497: stdout chunk (state=3): >>><<< 13731 1727203844.13509: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203844.13512: _low_level_execute_command(): starting 13731 1727203844.13517: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/AnsiballZ_command.py && sleep 0' 13731 1727203844.13927: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203844.13931: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.13933: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203844.13935: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.13984: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203844.13987: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203844.14040: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203844.31295: stdout chunk (state=3): >>> {"changed": true, "stdout": "bond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "start": "2024-09-24 14:50:44.291563", "end": "2024-09-24 14:50:44.312425", "delta": "0:00:00.020862", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203844.32840: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203844.32845: stdout chunk (state=3): >>><<< 13731 1727203844.33082: stderr chunk (state=3): >>><<< 13731 1727203844.33087: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "bond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection ", "stderr": "", "rc": 0, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "start": "2024-09-24 14:50:44.291563", "end": "2024-09-24 14:50:44.312425", "delta": "0:00:00.020862", "msg": "", "invocation": {"module_args": {"_raw_params": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203844.33095: done with _execute_module (ansible.legacy.command, {'_raw_params': 'nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203844.33099: _low_level_execute_command(): starting 13731 1727203844.33101: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203844.0126579-15109-201459069444535/ > /dev/null 2>&1 && sleep 0' 13731 1727203844.33566: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203844.33589: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203844.33597: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.33645: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203844.33704: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203844.33714: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203844.33781: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203844.33789: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203844.35653: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203844.35657: stdout chunk (state=3): >>><<< 13731 1727203844.35659: stderr chunk (state=3): >>><<< 13731 1727203844.35937: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203844.35940: handler run complete 13731 1727203844.35943: Evaluated conditional (False): False 13731 1727203844.35945: attempt loop complete, returning result 13731 1727203844.35951: _execute() done 13731 1727203844.35957: dumping result to json 13731 1727203844.35968: done dumping result, returning 13731 1727203844.35988: done running TaskExecutor() for managed-node3/TASK: Get NM profile info [028d2410-947f-82dc-c122-0000000005b7] 13731 1727203844.35997: sending task result for task 028d2410-947f-82dc-c122-0000000005b7 ok: [managed-node3] => { "changed": false, "cmd": "nmcli -f NAME,FILENAME connection show |grep bond0.1 | grep /etc", "delta": "0:00:00.020862", "end": "2024-09-24 14:50:44.312425", "rc": 0, "start": "2024-09-24 14:50:44.291563" } STDOUT: bond0.1 /etc/NetworkManager/system-connections/bond0.1.nmconnection 13731 1727203844.36224: no more pending results, returning what we have 13731 1727203844.36229: results queue empty 13731 1727203844.36229: checking for any_errors_fatal 13731 1727203844.36242: done checking for any_errors_fatal 13731 1727203844.36243: checking for max_fail_percentage 13731 1727203844.36245: done checking for max_fail_percentage 13731 1727203844.36246: checking to see if all hosts have failed and the running result is not ok 13731 1727203844.36246: done checking to see if all hosts have failed 13731 1727203844.36247: getting the remaining hosts for this loop 13731 1727203844.36249: done getting the remaining hosts for this loop 13731 1727203844.36254: getting the next task for host managed-node3 13731 1727203844.36266: done getting next task for host managed-node3 13731 1727203844.36270: ^ task is: TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13731 1727203844.36277: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203844.36282: getting variables 13731 1727203844.36283: in VariableManager get_vars() 13731 1727203844.36431: Calling all_inventory to load vars for managed-node3 13731 1727203844.36434: Calling groups_inventory to load vars for managed-node3 13731 1727203844.36438: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203844.36482: done sending task result for task 028d2410-947f-82dc-c122-0000000005b7 13731 1727203844.36489: WORKER PROCESS EXITING 13731 1727203844.36500: Calling all_plugins_play to load vars for managed-node3 13731 1727203844.36601: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203844.36606: Calling groups_plugins_play to load vars for managed-node3 13731 1727203844.38627: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203844.40438: done with get_vars() 13731 1727203844.40465: done getting variables 13731 1727203844.40528: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set NM profile exist flag and ansible_managed flag true based on the nmcli output] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:35 Tuesday 24 September 2024 14:50:44 -0400 (0:00:00.453) 0:00:22.636 ***** 13731 1727203844.40564: entering _queue_task() for managed-node3/set_fact 13731 1727203844.40916: worker is 1 (out of 1 available) 13731 1727203844.40927: exiting _queue_task() for managed-node3/set_fact 13731 1727203844.40939: done queuing things up, now waiting for results queue to drain 13731 1727203844.40941: waiting for pending results... 13731 1727203844.41226: running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output 13731 1727203844.41365: in run() - task 028d2410-947f-82dc-c122-0000000005b8 13731 1727203844.41487: variable 'ansible_search_path' from source: unknown 13731 1727203844.41491: variable 'ansible_search_path' from source: unknown 13731 1727203844.41494: calling self._execute() 13731 1727203844.41528: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.41532: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.41541: variable 'omit' from source: magic vars 13731 1727203844.41921: variable 'ansible_distribution_major_version' from source: facts 13731 1727203844.41931: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203844.42074: variable 'nm_profile_exists' from source: set_fact 13731 1727203844.42088: Evaluated conditional (nm_profile_exists.rc == 0): True 13731 1727203844.42093: variable 'omit' from source: magic vars 13731 1727203844.42160: variable 'omit' from source: magic vars 13731 1727203844.42195: variable 'omit' from source: magic vars 13731 1727203844.42236: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203844.42280: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203844.42299: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203844.42316: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203844.42328: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203844.42368: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203844.42371: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.42465: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.42482: Set connection var ansible_pipelining to False 13731 1727203844.42493: Set connection var ansible_shell_type to sh 13731 1727203844.42500: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203844.42505: Set connection var ansible_connection to ssh 13731 1727203844.42511: Set connection var ansible_shell_executable to /bin/sh 13731 1727203844.42516: Set connection var ansible_timeout to 10 13731 1727203844.42536: variable 'ansible_shell_executable' from source: unknown 13731 1727203844.42539: variable 'ansible_connection' from source: unknown 13731 1727203844.42542: variable 'ansible_module_compression' from source: unknown 13731 1727203844.42544: variable 'ansible_shell_type' from source: unknown 13731 1727203844.42546: variable 'ansible_shell_executable' from source: unknown 13731 1727203844.42548: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.42552: variable 'ansible_pipelining' from source: unknown 13731 1727203844.42555: variable 'ansible_timeout' from source: unknown 13731 1727203844.42558: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.42688: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203844.42699: variable 'omit' from source: magic vars 13731 1727203844.42793: starting attempt loop 13731 1727203844.42795: running the handler 13731 1727203844.42797: handler run complete 13731 1727203844.42799: attempt loop complete, returning result 13731 1727203844.42801: _execute() done 13731 1727203844.42802: dumping result to json 13731 1727203844.42804: done dumping result, returning 13731 1727203844.42806: done running TaskExecutor() for managed-node3/TASK: Set NM profile exist flag and ansible_managed flag true based on the nmcli output [028d2410-947f-82dc-c122-0000000005b8] 13731 1727203844.42807: sending task result for task 028d2410-947f-82dc-c122-0000000005b8 13731 1727203844.42869: done sending task result for task 028d2410-947f-82dc-c122-0000000005b8 13731 1727203844.42872: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "lsr_net_profile_ansible_managed": true, "lsr_net_profile_exists": true, "lsr_net_profile_fingerprint": true }, "changed": false } 13731 1727203844.42929: no more pending results, returning what we have 13731 1727203844.42933: results queue empty 13731 1727203844.42934: checking for any_errors_fatal 13731 1727203844.42944: done checking for any_errors_fatal 13731 1727203844.42945: checking for max_fail_percentage 13731 1727203844.42947: done checking for max_fail_percentage 13731 1727203844.42948: checking to see if all hosts have failed and the running result is not ok 13731 1727203844.42949: done checking to see if all hosts have failed 13731 1727203844.42950: getting the remaining hosts for this loop 13731 1727203844.42952: done getting the remaining hosts for this loop 13731 1727203844.42956: getting the next task for host managed-node3 13731 1727203844.42971: done getting next task for host managed-node3 13731 1727203844.42977: ^ task is: TASK: Get the ansible_managed comment in ifcfg-{{ profile }} 13731 1727203844.42985: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203844.42989: getting variables 13731 1727203844.42991: in VariableManager get_vars() 13731 1727203844.43025: Calling all_inventory to load vars for managed-node3 13731 1727203844.43028: Calling groups_inventory to load vars for managed-node3 13731 1727203844.43032: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203844.43044: Calling all_plugins_play to load vars for managed-node3 13731 1727203844.43046: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203844.43050: Calling groups_plugins_play to load vars for managed-node3 13731 1727203844.44608: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203844.46117: done with get_vars() 13731 1727203844.46147: done getting variables 13731 1727203844.46209: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203844.46327: variable 'profile' from source: include params 13731 1727203844.46331: variable 'bond_port_profile' from source: include params 13731 1727203844.46393: variable 'bond_port_profile' from source: include params TASK [Get the ansible_managed comment in ifcfg-bond0.1] ************************ task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:49 Tuesday 24 September 2024 14:50:44 -0400 (0:00:00.058) 0:00:22.695 ***** 13731 1727203844.46425: entering _queue_task() for managed-node3/command 13731 1727203844.46764: worker is 1 (out of 1 available) 13731 1727203844.46779: exiting _queue_task() for managed-node3/command 13731 1727203844.46792: done queuing things up, now waiting for results queue to drain 13731 1727203844.46793: waiting for pending results... 13731 1727203844.47230: running TaskExecutor() for managed-node3/TASK: Get the ansible_managed comment in ifcfg-bond0.1 13731 1727203844.47246: in run() - task 028d2410-947f-82dc-c122-0000000005ba 13731 1727203844.47261: variable 'ansible_search_path' from source: unknown 13731 1727203844.47264: variable 'ansible_search_path' from source: unknown 13731 1727203844.47310: calling self._execute() 13731 1727203844.47420: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.47424: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.47434: variable 'omit' from source: magic vars 13731 1727203844.47826: variable 'ansible_distribution_major_version' from source: facts 13731 1727203844.47842: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203844.47955: variable 'profile_stat' from source: set_fact 13731 1727203844.47977: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203844.47980: when evaluation is False, skipping this task 13731 1727203844.47983: _execute() done 13731 1727203844.47986: dumping result to json 13731 1727203844.47988: done dumping result, returning 13731 1727203844.47991: done running TaskExecutor() for managed-node3/TASK: Get the ansible_managed comment in ifcfg-bond0.1 [028d2410-947f-82dc-c122-0000000005ba] 13731 1727203844.47993: sending task result for task 028d2410-947f-82dc-c122-0000000005ba 13731 1727203844.48140: done sending task result for task 028d2410-947f-82dc-c122-0000000005ba 13731 1727203844.48144: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203844.48208: no more pending results, returning what we have 13731 1727203844.48212: results queue empty 13731 1727203844.48213: checking for any_errors_fatal 13731 1727203844.48221: done checking for any_errors_fatal 13731 1727203844.48222: checking for max_fail_percentage 13731 1727203844.48223: done checking for max_fail_percentage 13731 1727203844.48224: checking to see if all hosts have failed and the running result is not ok 13731 1727203844.48225: done checking to see if all hosts have failed 13731 1727203844.48226: getting the remaining hosts for this loop 13731 1727203844.48227: done getting the remaining hosts for this loop 13731 1727203844.48231: getting the next task for host managed-node3 13731 1727203844.48240: done getting next task for host managed-node3 13731 1727203844.48243: ^ task is: TASK: Verify the ansible_managed comment in ifcfg-{{ profile }} 13731 1727203844.48250: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203844.48254: getting variables 13731 1727203844.48256: in VariableManager get_vars() 13731 1727203844.48292: Calling all_inventory to load vars for managed-node3 13731 1727203844.48295: Calling groups_inventory to load vars for managed-node3 13731 1727203844.48299: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203844.48312: Calling all_plugins_play to load vars for managed-node3 13731 1727203844.48315: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203844.48317: Calling groups_plugins_play to load vars for managed-node3 13731 1727203844.49952: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203844.53108: done with get_vars() 13731 1727203844.53136: done getting variables 13731 1727203844.53247: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203844.53624: variable 'profile' from source: include params 13731 1727203844.53629: variable 'bond_port_profile' from source: include params 13731 1727203844.53824: variable 'bond_port_profile' from source: include params TASK [Verify the ansible_managed comment in ifcfg-bond0.1] ********************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:56 Tuesday 24 September 2024 14:50:44 -0400 (0:00:00.074) 0:00:22.769 ***** 13731 1727203844.53862: entering _queue_task() for managed-node3/set_fact 13731 1727203844.54442: worker is 1 (out of 1 available) 13731 1727203844.54572: exiting _queue_task() for managed-node3/set_fact 13731 1727203844.54586: done queuing things up, now waiting for results queue to drain 13731 1727203844.54588: waiting for pending results... 13731 1727203844.54886: running TaskExecutor() for managed-node3/TASK: Verify the ansible_managed comment in ifcfg-bond0.1 13731 1727203844.54954: in run() - task 028d2410-947f-82dc-c122-0000000005bb 13731 1727203844.54989: variable 'ansible_search_path' from source: unknown 13731 1727203844.54997: variable 'ansible_search_path' from source: unknown 13731 1727203844.55041: calling self._execute() 13731 1727203844.55188: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.55206: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.55221: variable 'omit' from source: magic vars 13731 1727203844.55643: variable 'ansible_distribution_major_version' from source: facts 13731 1727203844.55664: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203844.55793: variable 'profile_stat' from source: set_fact 13731 1727203844.55810: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203844.55817: when evaluation is False, skipping this task 13731 1727203844.55823: _execute() done 13731 1727203844.55830: dumping result to json 13731 1727203844.55841: done dumping result, returning 13731 1727203844.55860: done running TaskExecutor() for managed-node3/TASK: Verify the ansible_managed comment in ifcfg-bond0.1 [028d2410-947f-82dc-c122-0000000005bb] 13731 1727203844.55952: sending task result for task 028d2410-947f-82dc-c122-0000000005bb 13731 1727203844.56027: done sending task result for task 028d2410-947f-82dc-c122-0000000005bb 13731 1727203844.56029: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203844.56216: no more pending results, returning what we have 13731 1727203844.56219: results queue empty 13731 1727203844.56220: checking for any_errors_fatal 13731 1727203844.56225: done checking for any_errors_fatal 13731 1727203844.56226: checking for max_fail_percentage 13731 1727203844.56227: done checking for max_fail_percentage 13731 1727203844.56228: checking to see if all hosts have failed and the running result is not ok 13731 1727203844.56229: done checking to see if all hosts have failed 13731 1727203844.56230: getting the remaining hosts for this loop 13731 1727203844.56231: done getting the remaining hosts for this loop 13731 1727203844.56234: getting the next task for host managed-node3 13731 1727203844.56241: done getting next task for host managed-node3 13731 1727203844.56243: ^ task is: TASK: Get the fingerprint comment in ifcfg-{{ profile }} 13731 1727203844.56249: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203844.56253: getting variables 13731 1727203844.56255: in VariableManager get_vars() 13731 1727203844.56282: Calling all_inventory to load vars for managed-node3 13731 1727203844.56285: Calling groups_inventory to load vars for managed-node3 13731 1727203844.56288: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203844.56297: Calling all_plugins_play to load vars for managed-node3 13731 1727203844.56300: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203844.56302: Calling groups_plugins_play to load vars for managed-node3 13731 1727203844.58142: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203844.59882: done with get_vars() 13731 1727203844.59906: done getting variables 13731 1727203844.59963: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203844.60081: variable 'profile' from source: include params 13731 1727203844.60085: variable 'bond_port_profile' from source: include params 13731 1727203844.60146: variable 'bond_port_profile' from source: include params TASK [Get the fingerprint comment in ifcfg-bond0.1] **************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:62 Tuesday 24 September 2024 14:50:44 -0400 (0:00:00.063) 0:00:22.832 ***** 13731 1727203844.60181: entering _queue_task() for managed-node3/command 13731 1727203844.60512: worker is 1 (out of 1 available) 13731 1727203844.60778: exiting _queue_task() for managed-node3/command 13731 1727203844.60787: done queuing things up, now waiting for results queue to drain 13731 1727203844.60789: waiting for pending results... 13731 1727203844.61007: running TaskExecutor() for managed-node3/TASK: Get the fingerprint comment in ifcfg-bond0.1 13731 1727203844.61019: in run() - task 028d2410-947f-82dc-c122-0000000005bc 13731 1727203844.61023: variable 'ansible_search_path' from source: unknown 13731 1727203844.61025: variable 'ansible_search_path' from source: unknown 13731 1727203844.61029: calling self._execute() 13731 1727203844.61086: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.61092: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.61102: variable 'omit' from source: magic vars 13731 1727203844.61682: variable 'ansible_distribution_major_version' from source: facts 13731 1727203844.61685: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203844.61689: variable 'profile_stat' from source: set_fact 13731 1727203844.61698: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203844.61701: when evaluation is False, skipping this task 13731 1727203844.61704: _execute() done 13731 1727203844.61707: dumping result to json 13731 1727203844.61710: done dumping result, returning 13731 1727203844.61771: done running TaskExecutor() for managed-node3/TASK: Get the fingerprint comment in ifcfg-bond0.1 [028d2410-947f-82dc-c122-0000000005bc] 13731 1727203844.61781: sending task result for task 028d2410-947f-82dc-c122-0000000005bc 13731 1727203844.61843: done sending task result for task 028d2410-947f-82dc-c122-0000000005bc 13731 1727203844.61846: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203844.61901: no more pending results, returning what we have 13731 1727203844.61906: results queue empty 13731 1727203844.61907: checking for any_errors_fatal 13731 1727203844.61915: done checking for any_errors_fatal 13731 1727203844.61916: checking for max_fail_percentage 13731 1727203844.61918: done checking for max_fail_percentage 13731 1727203844.61919: checking to see if all hosts have failed and the running result is not ok 13731 1727203844.61920: done checking to see if all hosts have failed 13731 1727203844.61920: getting the remaining hosts for this loop 13731 1727203844.61923: done getting the remaining hosts for this loop 13731 1727203844.61926: getting the next task for host managed-node3 13731 1727203844.61935: done getting next task for host managed-node3 13731 1727203844.61938: ^ task is: TASK: Verify the fingerprint comment in ifcfg-{{ profile }} 13731 1727203844.61945: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203844.61949: getting variables 13731 1727203844.61951: in VariableManager get_vars() 13731 1727203844.61986: Calling all_inventory to load vars for managed-node3 13731 1727203844.61989: Calling groups_inventory to load vars for managed-node3 13731 1727203844.61993: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203844.62006: Calling all_plugins_play to load vars for managed-node3 13731 1727203844.62009: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203844.62011: Calling groups_plugins_play to load vars for managed-node3 13731 1727203844.63772: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203844.67450: done with get_vars() 13731 1727203844.67484: done getting variables 13731 1727203844.67540: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203844.67950: variable 'profile' from source: include params 13731 1727203844.67954: variable 'bond_port_profile' from source: include params 13731 1727203844.68014: variable 'bond_port_profile' from source: include params TASK [Verify the fingerprint comment in ifcfg-bond0.1] ************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_profile_stat.yml:69 Tuesday 24 September 2024 14:50:44 -0400 (0:00:00.078) 0:00:22.911 ***** 13731 1727203844.68047: entering _queue_task() for managed-node3/set_fact 13731 1727203844.68806: worker is 1 (out of 1 available) 13731 1727203844.68821: exiting _queue_task() for managed-node3/set_fact 13731 1727203844.68833: done queuing things up, now waiting for results queue to drain 13731 1727203844.68835: waiting for pending results... 13731 1727203844.69528: running TaskExecutor() for managed-node3/TASK: Verify the fingerprint comment in ifcfg-bond0.1 13731 1727203844.69652: in run() - task 028d2410-947f-82dc-c122-0000000005bd 13731 1727203844.69784: variable 'ansible_search_path' from source: unknown 13731 1727203844.69787: variable 'ansible_search_path' from source: unknown 13731 1727203844.69790: calling self._execute() 13731 1727203844.70004: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.70008: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.70055: variable 'omit' from source: magic vars 13731 1727203844.70734: variable 'ansible_distribution_major_version' from source: facts 13731 1727203844.70745: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203844.71031: variable 'profile_stat' from source: set_fact 13731 1727203844.71034: Evaluated conditional (profile_stat.stat.exists): False 13731 1727203844.71036: when evaluation is False, skipping this task 13731 1727203844.71038: _execute() done 13731 1727203844.71040: dumping result to json 13731 1727203844.71042: done dumping result, returning 13731 1727203844.71044: done running TaskExecutor() for managed-node3/TASK: Verify the fingerprint comment in ifcfg-bond0.1 [028d2410-947f-82dc-c122-0000000005bd] 13731 1727203844.71045: sending task result for task 028d2410-947f-82dc-c122-0000000005bd 13731 1727203844.71311: done sending task result for task 028d2410-947f-82dc-c122-0000000005bd 13731 1727203844.71314: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "profile_stat.stat.exists", "skip_reason": "Conditional result was False" } 13731 1727203844.71393: no more pending results, returning what we have 13731 1727203844.71398: results queue empty 13731 1727203844.71398: checking for any_errors_fatal 13731 1727203844.71406: done checking for any_errors_fatal 13731 1727203844.71406: checking for max_fail_percentage 13731 1727203844.71409: done checking for max_fail_percentage 13731 1727203844.71410: checking to see if all hosts have failed and the running result is not ok 13731 1727203844.71410: done checking to see if all hosts have failed 13731 1727203844.71411: getting the remaining hosts for this loop 13731 1727203844.71413: done getting the remaining hosts for this loop 13731 1727203844.71417: getting the next task for host managed-node3 13731 1727203844.71428: done getting next task for host managed-node3 13731 1727203844.71431: ^ task is: TASK: Assert that the profile is present - '{{ profile }}' 13731 1727203844.71438: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203844.71442: getting variables 13731 1727203844.71444: in VariableManager get_vars() 13731 1727203844.71478: Calling all_inventory to load vars for managed-node3 13731 1727203844.71481: Calling groups_inventory to load vars for managed-node3 13731 1727203844.71486: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203844.71498: Calling all_plugins_play to load vars for managed-node3 13731 1727203844.71501: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203844.71504: Calling groups_plugins_play to load vars for managed-node3 13731 1727203844.73636: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203844.77343: done with get_vars() 13731 1727203844.77371: done getting variables 13731 1727203844.77437: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203844.77789: variable 'profile' from source: include params 13731 1727203844.77793: variable 'bond_port_profile' from source: include params 13731 1727203844.78001: variable 'bond_port_profile' from source: include params TASK [Assert that the profile is present - 'bond0.1'] ************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:5 Tuesday 24 September 2024 14:50:44 -0400 (0:00:00.100) 0:00:23.012 ***** 13731 1727203844.78117: entering _queue_task() for managed-node3/assert 13731 1727203844.78870: worker is 1 (out of 1 available) 13731 1727203844.78938: exiting _queue_task() for managed-node3/assert 13731 1727203844.78947: done queuing things up, now waiting for results queue to drain 13731 1727203844.78949: waiting for pending results... 13731 1727203844.79997: running TaskExecutor() for managed-node3/TASK: Assert that the profile is present - 'bond0.1' 13731 1727203844.80227: in run() - task 028d2410-947f-82dc-c122-0000000004e8 13731 1727203844.80231: variable 'ansible_search_path' from source: unknown 13731 1727203844.80233: variable 'ansible_search_path' from source: unknown 13731 1727203844.80662: calling self._execute() 13731 1727203844.80667: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.80681: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.80694: variable 'omit' from source: magic vars 13731 1727203844.81873: variable 'ansible_distribution_major_version' from source: facts 13731 1727203844.81896: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203844.81907: variable 'omit' from source: magic vars 13731 1727203844.82001: variable 'omit' from source: magic vars 13731 1727203844.82384: variable 'profile' from source: include params 13731 1727203844.82493: variable 'bond_port_profile' from source: include params 13731 1727203844.82560: variable 'bond_port_profile' from source: include params 13731 1727203844.82730: variable 'omit' from source: magic vars 13731 1727203844.82777: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203844.82887: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203844.82911: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203844.82937: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203844.83145: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203844.83148: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203844.83151: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.83153: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.83313: Set connection var ansible_pipelining to False 13731 1727203844.83372: Set connection var ansible_shell_type to sh 13731 1727203844.83385: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203844.83395: Set connection var ansible_connection to ssh 13731 1727203844.83404: Set connection var ansible_shell_executable to /bin/sh 13731 1727203844.83413: Set connection var ansible_timeout to 10 13731 1727203844.83443: variable 'ansible_shell_executable' from source: unknown 13731 1727203844.83680: variable 'ansible_connection' from source: unknown 13731 1727203844.83685: variable 'ansible_module_compression' from source: unknown 13731 1727203844.83687: variable 'ansible_shell_type' from source: unknown 13731 1727203844.83689: variable 'ansible_shell_executable' from source: unknown 13731 1727203844.83691: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.83693: variable 'ansible_pipelining' from source: unknown 13731 1727203844.83695: variable 'ansible_timeout' from source: unknown 13731 1727203844.83697: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.83771: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203844.84033: variable 'omit' from source: magic vars 13731 1727203844.84037: starting attempt loop 13731 1727203844.84039: running the handler 13731 1727203844.84115: variable 'lsr_net_profile_exists' from source: set_fact 13731 1727203844.84360: Evaluated conditional (lsr_net_profile_exists): True 13731 1727203844.84363: handler run complete 13731 1727203844.84366: attempt loop complete, returning result 13731 1727203844.84368: _execute() done 13731 1727203844.84370: dumping result to json 13731 1727203844.84372: done dumping result, returning 13731 1727203844.84374: done running TaskExecutor() for managed-node3/TASK: Assert that the profile is present - 'bond0.1' [028d2410-947f-82dc-c122-0000000004e8] 13731 1727203844.84379: sending task result for task 028d2410-947f-82dc-c122-0000000004e8 13731 1727203844.84449: done sending task result for task 028d2410-947f-82dc-c122-0000000004e8 13731 1727203844.84452: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203844.84607: no more pending results, returning what we have 13731 1727203844.84611: results queue empty 13731 1727203844.84612: checking for any_errors_fatal 13731 1727203844.84623: done checking for any_errors_fatal 13731 1727203844.84624: checking for max_fail_percentage 13731 1727203844.84625: done checking for max_fail_percentage 13731 1727203844.84626: checking to see if all hosts have failed and the running result is not ok 13731 1727203844.84627: done checking to see if all hosts have failed 13731 1727203844.84627: getting the remaining hosts for this loop 13731 1727203844.84629: done getting the remaining hosts for this loop 13731 1727203844.84632: getting the next task for host managed-node3 13731 1727203844.84639: done getting next task for host managed-node3 13731 1727203844.84642: ^ task is: TASK: Assert that the ansible managed comment is present in '{{ profile }}' 13731 1727203844.84647: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203844.84651: getting variables 13731 1727203844.84652: in VariableManager get_vars() 13731 1727203844.84836: Calling all_inventory to load vars for managed-node3 13731 1727203844.84839: Calling groups_inventory to load vars for managed-node3 13731 1727203844.84843: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203844.84854: Calling all_plugins_play to load vars for managed-node3 13731 1727203844.84856: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203844.84858: Calling groups_plugins_play to load vars for managed-node3 13731 1727203844.88242: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203844.91792: done with get_vars() 13731 1727203844.91822: done getting variables 13731 1727203844.91898: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203844.92219: variable 'profile' from source: include params 13731 1727203844.92223: variable 'bond_port_profile' from source: include params 13731 1727203844.92399: variable 'bond_port_profile' from source: include params TASK [Assert that the ansible managed comment is present in 'bond0.1'] ********* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:10 Tuesday 24 September 2024 14:50:44 -0400 (0:00:00.143) 0:00:23.155 ***** 13731 1727203844.92432: entering _queue_task() for managed-node3/assert 13731 1727203844.93195: worker is 1 (out of 1 available) 13731 1727203844.93208: exiting _queue_task() for managed-node3/assert 13731 1727203844.93220: done queuing things up, now waiting for results queue to drain 13731 1727203844.93222: waiting for pending results... 13731 1727203844.93893: running TaskExecutor() for managed-node3/TASK: Assert that the ansible managed comment is present in 'bond0.1' 13731 1727203844.93905: in run() - task 028d2410-947f-82dc-c122-0000000004e9 13731 1727203844.93909: variable 'ansible_search_path' from source: unknown 13731 1727203844.93911: variable 'ansible_search_path' from source: unknown 13731 1727203844.93923: calling self._execute() 13731 1727203844.94082: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.94085: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.94088: variable 'omit' from source: magic vars 13731 1727203844.94480: variable 'ansible_distribution_major_version' from source: facts 13731 1727203844.94484: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203844.94486: variable 'omit' from source: magic vars 13731 1727203844.94489: variable 'omit' from source: magic vars 13731 1727203844.94581: variable 'profile' from source: include params 13731 1727203844.94585: variable 'bond_port_profile' from source: include params 13731 1727203844.94645: variable 'bond_port_profile' from source: include params 13731 1727203844.94664: variable 'omit' from source: magic vars 13731 1727203844.94712: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203844.94847: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203844.94851: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203844.94853: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203844.94856: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203844.94858: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203844.94860: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.94862: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.94958: Set connection var ansible_pipelining to False 13731 1727203844.94961: Set connection var ansible_shell_type to sh 13731 1727203844.94999: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203844.95002: Set connection var ansible_connection to ssh 13731 1727203844.95004: Set connection var ansible_shell_executable to /bin/sh 13731 1727203844.95006: Set connection var ansible_timeout to 10 13731 1727203844.95016: variable 'ansible_shell_executable' from source: unknown 13731 1727203844.95018: variable 'ansible_connection' from source: unknown 13731 1727203844.95021: variable 'ansible_module_compression' from source: unknown 13731 1727203844.95023: variable 'ansible_shell_type' from source: unknown 13731 1727203844.95025: variable 'ansible_shell_executable' from source: unknown 13731 1727203844.95107: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203844.95110: variable 'ansible_pipelining' from source: unknown 13731 1727203844.95113: variable 'ansible_timeout' from source: unknown 13731 1727203844.95116: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203844.95437: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203844.95441: variable 'omit' from source: magic vars 13731 1727203844.95443: starting attempt loop 13731 1727203844.95446: running the handler 13731 1727203844.95599: variable 'lsr_net_profile_ansible_managed' from source: set_fact 13731 1727203844.95602: Evaluated conditional (lsr_net_profile_ansible_managed): True 13731 1727203844.95609: handler run complete 13731 1727203844.95624: attempt loop complete, returning result 13731 1727203844.95627: _execute() done 13731 1727203844.95629: dumping result to json 13731 1727203844.95632: done dumping result, returning 13731 1727203844.95639: done running TaskExecutor() for managed-node3/TASK: Assert that the ansible managed comment is present in 'bond0.1' [028d2410-947f-82dc-c122-0000000004e9] 13731 1727203844.95648: sending task result for task 028d2410-947f-82dc-c122-0000000004e9 13731 1727203844.95964: done sending task result for task 028d2410-947f-82dc-c122-0000000004e9 13731 1727203844.95967: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203844.96033: no more pending results, returning what we have 13731 1727203844.96036: results queue empty 13731 1727203844.96037: checking for any_errors_fatal 13731 1727203844.96045: done checking for any_errors_fatal 13731 1727203844.96046: checking for max_fail_percentage 13731 1727203844.96048: done checking for max_fail_percentage 13731 1727203844.96049: checking to see if all hosts have failed and the running result is not ok 13731 1727203844.96049: done checking to see if all hosts have failed 13731 1727203844.96051: getting the remaining hosts for this loop 13731 1727203844.96052: done getting the remaining hosts for this loop 13731 1727203844.96056: getting the next task for host managed-node3 13731 1727203844.96067: done getting next task for host managed-node3 13731 1727203844.96069: ^ task is: TASK: Assert that the fingerprint comment is present in {{ profile }} 13731 1727203844.96074: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203844.96082: getting variables 13731 1727203844.96088: in VariableManager get_vars() 13731 1727203844.96121: Calling all_inventory to load vars for managed-node3 13731 1727203844.96124: Calling groups_inventory to load vars for managed-node3 13731 1727203844.96128: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203844.96139: Calling all_plugins_play to load vars for managed-node3 13731 1727203844.96142: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203844.96145: Calling groups_plugins_play to load vars for managed-node3 13731 1727203844.99436: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203845.02926: done with get_vars() 13731 1727203845.02955: done getting variables 13731 1727203845.03044: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203845.03288: variable 'profile' from source: include params 13731 1727203845.03292: variable 'bond_port_profile' from source: include params 13731 1727203845.03457: variable 'bond_port_profile' from source: include params TASK [Assert that the fingerprint comment is present in bond0.1] *************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_profile_present.yml:15 Tuesday 24 September 2024 14:50:45 -0400 (0:00:00.111) 0:00:23.266 ***** 13731 1727203845.03552: entering _queue_task() for managed-node3/assert 13731 1727203845.04277: worker is 1 (out of 1 available) 13731 1727203845.04290: exiting _queue_task() for managed-node3/assert 13731 1727203845.04379: done queuing things up, now waiting for results queue to drain 13731 1727203845.04382: waiting for pending results... 13731 1727203845.05170: running TaskExecutor() for managed-node3/TASK: Assert that the fingerprint comment is present in bond0.1 13731 1727203845.05249: in run() - task 028d2410-947f-82dc-c122-0000000004ea 13731 1727203845.05265: variable 'ansible_search_path' from source: unknown 13731 1727203845.05268: variable 'ansible_search_path' from source: unknown 13731 1727203845.05306: calling self._execute() 13731 1727203845.05521: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.05524: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.05535: variable 'omit' from source: magic vars 13731 1727203845.06098: variable 'ansible_distribution_major_version' from source: facts 13731 1727203845.06116: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203845.06120: variable 'omit' from source: magic vars 13731 1727203845.06174: variable 'omit' from source: magic vars 13731 1727203845.06287: variable 'profile' from source: include params 13731 1727203845.06293: variable 'bond_port_profile' from source: include params 13731 1727203845.06357: variable 'bond_port_profile' from source: include params 13731 1727203845.06381: variable 'omit' from source: magic vars 13731 1727203845.06420: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203845.06459: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203845.06483: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203845.06500: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203845.06512: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203845.06547: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203845.06550: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.06553: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.06658: Set connection var ansible_pipelining to False 13731 1727203845.06661: Set connection var ansible_shell_type to sh 13731 1727203845.06673: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203845.06677: Set connection var ansible_connection to ssh 13731 1727203845.06682: Set connection var ansible_shell_executable to /bin/sh 13731 1727203845.06687: Set connection var ansible_timeout to 10 13731 1727203845.06710: variable 'ansible_shell_executable' from source: unknown 13731 1727203845.06713: variable 'ansible_connection' from source: unknown 13731 1727203845.06715: variable 'ansible_module_compression' from source: unknown 13731 1727203845.06718: variable 'ansible_shell_type' from source: unknown 13731 1727203845.06720: variable 'ansible_shell_executable' from source: unknown 13731 1727203845.06722: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.06726: variable 'ansible_pipelining' from source: unknown 13731 1727203845.06729: variable 'ansible_timeout' from source: unknown 13731 1727203845.06731: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.06895: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203845.07064: variable 'omit' from source: magic vars 13731 1727203845.07084: starting attempt loop 13731 1727203845.07087: running the handler 13731 1727203845.07302: variable 'lsr_net_profile_fingerprint' from source: set_fact 13731 1727203845.07305: Evaluated conditional (lsr_net_profile_fingerprint): True 13731 1727203845.07325: handler run complete 13731 1727203845.07331: attempt loop complete, returning result 13731 1727203845.07435: _execute() done 13731 1727203845.07438: dumping result to json 13731 1727203845.07442: done dumping result, returning 13731 1727203845.07444: done running TaskExecutor() for managed-node3/TASK: Assert that the fingerprint comment is present in bond0.1 [028d2410-947f-82dc-c122-0000000004ea] 13731 1727203845.07447: sending task result for task 028d2410-947f-82dc-c122-0000000004ea 13731 1727203845.07515: done sending task result for task 028d2410-947f-82dc-c122-0000000004ea 13731 1727203845.07518: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203845.07566: no more pending results, returning what we have 13731 1727203845.07570: results queue empty 13731 1727203845.07571: checking for any_errors_fatal 13731 1727203845.07580: done checking for any_errors_fatal 13731 1727203845.07581: checking for max_fail_percentage 13731 1727203845.07583: done checking for max_fail_percentage 13731 1727203845.07584: checking to see if all hosts have failed and the running result is not ok 13731 1727203845.07585: done checking to see if all hosts have failed 13731 1727203845.07586: getting the remaining hosts for this loop 13731 1727203845.07588: done getting the remaining hosts for this loop 13731 1727203845.07591: getting the next task for host managed-node3 13731 1727203845.07604: done getting next task for host managed-node3 13731 1727203845.07608: ^ task is: TASK: ** TEST check bond settings 13731 1727203845.07612: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203845.07617: getting variables 13731 1727203845.07619: in VariableManager get_vars() 13731 1727203845.07653: Calling all_inventory to load vars for managed-node3 13731 1727203845.07656: Calling groups_inventory to load vars for managed-node3 13731 1727203845.07660: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203845.07671: Calling all_plugins_play to load vars for managed-node3 13731 1727203845.07674: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203845.07783: Calling groups_plugins_play to load vars for managed-node3 13731 1727203845.09788: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203845.12552: done with get_vars() 13731 1727203845.12589: done getting variables 13731 1727203845.12655: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [** TEST check bond settings] ********************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml:3 Tuesday 24 September 2024 14:50:45 -0400 (0:00:00.091) 0:00:23.357 ***** 13731 1727203845.12700: entering _queue_task() for managed-node3/command 13731 1727203845.13065: worker is 1 (out of 1 available) 13731 1727203845.13080: exiting _queue_task() for managed-node3/command 13731 1727203845.13205: done queuing things up, now waiting for results queue to drain 13731 1727203845.13207: waiting for pending results... 13731 1727203845.13547: running TaskExecutor() for managed-node3/TASK: ** TEST check bond settings 13731 1727203845.13556: in run() - task 028d2410-947f-82dc-c122-000000000400 13731 1727203845.13564: variable 'ansible_search_path' from source: unknown 13731 1727203845.13568: variable 'ansible_search_path' from source: unknown 13731 1727203845.13709: variable 'bond_options_to_assert' from source: play vars 13731 1727203845.13804: variable 'bond_options_to_assert' from source: play vars 13731 1727203845.13999: variable 'omit' from source: magic vars 13731 1727203845.14186: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.14189: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.14191: variable 'omit' from source: magic vars 13731 1727203845.14545: variable 'ansible_distribution_major_version' from source: facts 13731 1727203845.14582: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203845.14586: variable 'omit' from source: magic vars 13731 1727203845.14791: variable 'omit' from source: magic vars 13731 1727203845.15030: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203845.18447: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203845.18602: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203845.18606: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203845.18609: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203845.18714: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203845.18723: variable 'controller_device' from source: play vars 13731 1727203845.18726: variable 'bond_opt' from source: unknown 13731 1727203845.18748: variable 'omit' from source: magic vars 13731 1727203845.18899: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203845.19111: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203845.19114: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203845.19117: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203845.19119: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203845.19121: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203845.19184: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.19193: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.19508: Set connection var ansible_pipelining to False 13731 1727203845.19523: Set connection var ansible_shell_type to sh 13731 1727203845.19530: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203845.19598: Set connection var ansible_connection to ssh 13731 1727203845.19607: Set connection var ansible_shell_executable to /bin/sh 13731 1727203845.19614: Set connection var ansible_timeout to 10 13731 1727203845.19753: variable 'ansible_shell_executable' from source: unknown 13731 1727203845.19756: variable 'ansible_connection' from source: unknown 13731 1727203845.19759: variable 'ansible_module_compression' from source: unknown 13731 1727203845.19855: variable 'ansible_shell_type' from source: unknown 13731 1727203845.19858: variable 'ansible_shell_executable' from source: unknown 13731 1727203845.19863: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.19867: variable 'ansible_pipelining' from source: unknown 13731 1727203845.19870: variable 'ansible_timeout' from source: unknown 13731 1727203845.19911: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.20266: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203845.20274: variable 'omit' from source: magic vars 13731 1727203845.20408: starting attempt loop 13731 1727203845.20411: running the handler 13731 1727203845.20433: _low_level_execute_command(): starting 13731 1727203845.20439: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203845.22023: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203845.22030: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203845.22108: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203845.22114: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.22282: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.22369: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.22409: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.24072: stdout chunk (state=3): >>>/root <<< 13731 1727203845.24253: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.24257: stdout chunk (state=3): >>><<< 13731 1727203845.24260: stderr chunk (state=3): >>><<< 13731 1727203845.24284: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203845.24383: _low_level_execute_command(): starting 13731 1727203845.24387: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783 `" && echo ansible-tmp-1727203845.2429981-15163-218588534625783="` echo /root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783 `" ) && sleep 0' 13731 1727203845.24958: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203845.24980: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203845.24995: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203845.25022: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203845.25139: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.25200: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.25267: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.27112: stdout chunk (state=3): >>>ansible-tmp-1727203845.2429981-15163-218588534625783=/root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783 <<< 13731 1727203845.27301: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.27305: stdout chunk (state=3): >>><<< 13731 1727203845.27307: stderr chunk (state=3): >>><<< 13731 1727203845.27382: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203845.2429981-15163-218588534625783=/root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203845.27385: variable 'ansible_module_compression' from source: unknown 13731 1727203845.27435: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203845.27482: variable 'ansible_facts' from source: unknown 13731 1727203845.27605: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/AnsiballZ_command.py 13731 1727203845.27913: Sending initial data 13731 1727203845.27916: Sent initial data (156 bytes) 13731 1727203845.28915: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.28950: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203845.28957: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.28978: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.29041: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.30590: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203845.30647: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203845.30732: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpxldj2r0b /root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/AnsiballZ_command.py <<< 13731 1727203845.30736: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/AnsiballZ_command.py" <<< 13731 1727203845.30767: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpxldj2r0b" to remote "/root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/AnsiballZ_command.py" <<< 13731 1727203845.31932: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.31985: stderr chunk (state=3): >>><<< 13731 1727203845.31995: stdout chunk (state=3): >>><<< 13731 1727203845.32153: done transferring module to remote 13731 1727203845.32156: _low_level_execute_command(): starting 13731 1727203845.32159: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/ /root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/AnsiballZ_command.py && sleep 0' 13731 1727203845.33411: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.33500: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203845.33618: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.33766: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.33931: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.33938: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.35736: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.35740: stdout chunk (state=3): >>><<< 13731 1727203845.35743: stderr chunk (state=3): >>><<< 13731 1727203845.35868: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203845.35871: _low_level_execute_command(): starting 13731 1727203845.35874: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/AnsiballZ_command.py && sleep 0' 13731 1727203845.36678: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203845.36781: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203845.36806: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203845.37006: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203845.37033: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.37057: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.37160: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.52570: stdout chunk (state=3): >>> {"changed": true, "stdout": "802.3ad 4", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/mode"], "start": "2024-09-24 14:50:45.521516", "end": "2024-09-24 14:50:45.524669", "delta": "0:00:00.003153", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/mode", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203845.54582: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203845.54586: stdout chunk (state=3): >>><<< 13731 1727203845.54589: stderr chunk (state=3): >>><<< 13731 1727203845.54591: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "802.3ad 4", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/mode"], "start": "2024-09-24 14:50:45.521516", "end": "2024-09-24 14:50:45.524669", "delta": "0:00:00.003153", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/mode", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203845.54594: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/mode', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203845.54596: _low_level_execute_command(): starting 13731 1727203845.54598: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203845.2429981-15163-218588534625783/ > /dev/null 2>&1 && sleep 0' 13731 1727203845.56054: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.56066: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.56217: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.56520: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.58453: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.58457: stderr chunk (state=3): >>><<< 13731 1727203845.58460: stdout chunk (state=3): >>><<< 13731 1727203845.58591: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203845.58595: handler run complete 13731 1727203845.58597: Evaluated conditional (False): False 13731 1727203845.58850: variable 'bond_opt' from source: unknown 13731 1727203845.58856: variable 'result' from source: unknown 13731 1727203845.58871: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203845.58884: attempt loop complete, returning result 13731 1727203845.58903: variable 'bond_opt' from source: unknown 13731 1727203845.59153: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'mode', 'value': '802.3ad'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "mode", "value": "802.3ad" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/mode" ], "delta": "0:00:00.003153", "end": "2024-09-24 14:50:45.524669", "rc": 0, "start": "2024-09-24 14:50:45.521516" } STDOUT: 802.3ad 4 13731 1727203845.59733: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.59737: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.59740: variable 'omit' from source: magic vars 13731 1727203845.60280: variable 'ansible_distribution_major_version' from source: facts 13731 1727203845.60284: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203845.60286: variable 'omit' from source: magic vars 13731 1727203845.60289: variable 'omit' from source: magic vars 13731 1727203845.60670: variable 'controller_device' from source: play vars 13731 1727203845.60673: variable 'bond_opt' from source: unknown 13731 1727203845.60696: variable 'omit' from source: magic vars 13731 1727203845.60838: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203845.60847: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203845.60853: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203845.60870: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203845.60873: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.60877: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.61001: Set connection var ansible_pipelining to False 13731 1727203845.61004: Set connection var ansible_shell_type to sh 13731 1727203845.61011: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203845.61015: Set connection var ansible_connection to ssh 13731 1727203845.61181: Set connection var ansible_shell_executable to /bin/sh 13731 1727203845.61184: Set connection var ansible_timeout to 10 13731 1727203845.61215: variable 'ansible_shell_executable' from source: unknown 13731 1727203845.61218: variable 'ansible_connection' from source: unknown 13731 1727203845.61221: variable 'ansible_module_compression' from source: unknown 13731 1727203845.61223: variable 'ansible_shell_type' from source: unknown 13731 1727203845.61226: variable 'ansible_shell_executable' from source: unknown 13731 1727203845.61228: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.61232: variable 'ansible_pipelining' from source: unknown 13731 1727203845.61235: variable 'ansible_timeout' from source: unknown 13731 1727203845.61239: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.61681: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203845.61688: variable 'omit' from source: magic vars 13731 1727203845.61691: starting attempt loop 13731 1727203845.61693: running the handler 13731 1727203845.61695: _low_level_execute_command(): starting 13731 1727203845.61697: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203845.62968: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203845.63063: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.63187: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203845.63200: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.63209: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.63285: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.64901: stdout chunk (state=3): >>>/root <<< 13731 1727203845.65090: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.65094: stdout chunk (state=3): >>><<< 13731 1727203845.65100: stderr chunk (state=3): >>><<< 13731 1727203845.65147: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203845.65158: _low_level_execute_command(): starting 13731 1727203845.65166: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400 `" && echo ansible-tmp-1727203845.651475-15163-142234095850400="` echo /root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400 `" ) && sleep 0' 13731 1727203845.66641: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203845.66645: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203845.66774: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203845.66780: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.66944: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.68798: stdout chunk (state=3): >>>ansible-tmp-1727203845.651475-15163-142234095850400=/root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400 <<< 13731 1727203845.68867: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.68871: stderr chunk (state=3): >>><<< 13731 1727203845.68878: stdout chunk (state=3): >>><<< 13731 1727203845.68907: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203845.651475-15163-142234095850400=/root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203845.68982: variable 'ansible_module_compression' from source: unknown 13731 1727203845.68985: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203845.68993: variable 'ansible_facts' from source: unknown 13731 1727203845.69283: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/AnsiballZ_command.py 13731 1727203845.69515: Sending initial data 13731 1727203845.69525: Sent initial data (155 bytes) 13731 1727203845.70929: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.71035: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.71136: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.71251: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.72649: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 <<< 13731 1727203845.72679: stderr chunk (state=3): >>>debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203845.72710: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203845.72746: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp4gmig0vk /root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/AnsiballZ_command.py <<< 13731 1727203845.72828: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/AnsiballZ_command.py" <<< 13731 1727203845.72849: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp4gmig0vk" to remote "/root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/AnsiballZ_command.py" <<< 13731 1727203845.74114: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.74161: stderr chunk (state=3): >>><<< 13731 1727203845.74189: stdout chunk (state=3): >>><<< 13731 1727203845.74385: done transferring module to remote 13731 1727203845.74388: _low_level_execute_command(): starting 13731 1727203845.74390: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/ /root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/AnsiballZ_command.py && sleep 0' 13731 1727203845.75570: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203845.75611: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203845.75791: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203845.75828: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.75936: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.76004: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.77747: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.77773: stdout chunk (state=3): >>><<< 13731 1727203845.77978: stderr chunk (state=3): >>><<< 13731 1727203845.77983: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203845.77986: _low_level_execute_command(): starting 13731 1727203845.77989: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/AnsiballZ_command.py && sleep 0' 13731 1727203845.79296: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.79339: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.79361: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.79438: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.94520: stdout chunk (state=3): >>> {"changed": true, "stdout": "65535", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/ad_actor_sys_prio"], "start": "2024-09-24 14:50:45.941713", "end": "2024-09-24 14:50:45.944742", "delta": "0:00:00.003029", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/ad_actor_sys_prio", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203845.96007: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203845.96015: stderr chunk (state=3): >>><<< 13731 1727203845.96018: stdout chunk (state=3): >>><<< 13731 1727203845.96047: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "65535", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/ad_actor_sys_prio"], "start": "2024-09-24 14:50:45.941713", "end": "2024-09-24 14:50:45.944742", "delta": "0:00:00.003029", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/ad_actor_sys_prio", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203845.96085: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/ad_actor_sys_prio', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203845.96089: _low_level_execute_command(): starting 13731 1727203845.96095: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203845.651475-15163-142234095850400/ > /dev/null 2>&1 && sleep 0' 13731 1727203845.96801: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203845.96811: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203845.96823: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203845.96870: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203845.96996: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203845.98774: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203845.98800: stderr chunk (state=3): >>><<< 13731 1727203845.98803: stdout chunk (state=3): >>><<< 13731 1727203845.98815: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203845.98821: handler run complete 13731 1727203845.98836: Evaluated conditional (False): False 13731 1727203845.98945: variable 'bond_opt' from source: unknown 13731 1727203845.98949: variable 'result' from source: unknown 13731 1727203845.98960: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203845.98977: attempt loop complete, returning result 13731 1727203845.98992: variable 'bond_opt' from source: unknown 13731 1727203845.99038: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'ad_actor_sys_prio', 'value': '65535'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "ad_actor_sys_prio", "value": "65535" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/ad_actor_sys_prio" ], "delta": "0:00:00.003029", "end": "2024-09-24 14:50:45.944742", "rc": 0, "start": "2024-09-24 14:50:45.941713" } STDOUT: 65535 13731 1727203845.99169: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.99172: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.99174: variable 'omit' from source: magic vars 13731 1727203845.99257: variable 'ansible_distribution_major_version' from source: facts 13731 1727203845.99260: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203845.99267: variable 'omit' from source: magic vars 13731 1727203845.99280: variable 'omit' from source: magic vars 13731 1727203845.99390: variable 'controller_device' from source: play vars 13731 1727203845.99393: variable 'bond_opt' from source: unknown 13731 1727203845.99410: variable 'omit' from source: magic vars 13731 1727203845.99426: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203845.99432: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203845.99438: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203845.99448: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203845.99451: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.99454: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.99507: Set connection var ansible_pipelining to False 13731 1727203845.99512: Set connection var ansible_shell_type to sh 13731 1727203845.99514: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203845.99526: Set connection var ansible_connection to ssh 13731 1727203845.99528: Set connection var ansible_shell_executable to /bin/sh 13731 1727203845.99530: Set connection var ansible_timeout to 10 13731 1727203845.99543: variable 'ansible_shell_executable' from source: unknown 13731 1727203845.99546: variable 'ansible_connection' from source: unknown 13731 1727203845.99548: variable 'ansible_module_compression' from source: unknown 13731 1727203845.99550: variable 'ansible_shell_type' from source: unknown 13731 1727203845.99552: variable 'ansible_shell_executable' from source: unknown 13731 1727203845.99555: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203845.99559: variable 'ansible_pipelining' from source: unknown 13731 1727203845.99561: variable 'ansible_timeout' from source: unknown 13731 1727203845.99567: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203845.99635: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203845.99638: variable 'omit' from source: magic vars 13731 1727203845.99640: starting attempt loop 13731 1727203845.99642: running the handler 13731 1727203845.99649: _low_level_execute_command(): starting 13731 1727203845.99651: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203846.00199: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.00202: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.00205: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.00207: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.00228: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.00231: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.00294: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.01841: stdout chunk (state=3): >>>/root <<< 13731 1727203846.01977: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.01981: stdout chunk (state=3): >>><<< 13731 1727203846.01983: stderr chunk (state=3): >>><<< 13731 1727203846.02077: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.02085: _low_level_execute_command(): starting 13731 1727203846.02088: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019 `" && echo ansible-tmp-1727203846.0200353-15163-197084158304019="` echo /root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019 `" ) && sleep 0' 13731 1727203846.02538: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203846.02556: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203846.02559: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.02614: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.02618: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.02659: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.04494: stdout chunk (state=3): >>>ansible-tmp-1727203846.0200353-15163-197084158304019=/root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019 <<< 13731 1727203846.04658: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.04661: stdout chunk (state=3): >>><<< 13731 1727203846.04663: stderr chunk (state=3): >>><<< 13731 1727203846.04883: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203846.0200353-15163-197084158304019=/root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.04887: variable 'ansible_module_compression' from source: unknown 13731 1727203846.04889: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203846.04891: variable 'ansible_facts' from source: unknown 13731 1727203846.04893: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/AnsiballZ_command.py 13731 1727203846.05055: Sending initial data 13731 1727203846.05069: Sent initial data (156 bytes) 13731 1727203846.05615: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.05627: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203846.05642: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.05699: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.05735: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.05801: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.07264: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13731 1727203846.07280: stderr chunk (state=3): >>>debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203846.07306: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203846.07345: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpto5xmqgs /root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/AnsiballZ_command.py <<< 13731 1727203846.07352: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/AnsiballZ_command.py" <<< 13731 1727203846.07376: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpto5xmqgs" to remote "/root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/AnsiballZ_command.py" <<< 13731 1727203846.08043: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.08081: stderr chunk (state=3): >>><<< 13731 1727203846.08088: stdout chunk (state=3): >>><<< 13731 1727203846.08127: done transferring module to remote 13731 1727203846.08130: _low_level_execute_command(): starting 13731 1727203846.08142: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/ /root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/AnsiballZ_command.py && sleep 0' 13731 1727203846.08734: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203846.08761: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203846.08764: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.08829: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.08840: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.08913: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.10615: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.10647: stderr chunk (state=3): >>><<< 13731 1727203846.10650: stdout chunk (state=3): >>><<< 13731 1727203846.10681: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.10684: _low_level_execute_command(): starting 13731 1727203846.10690: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/AnsiballZ_command.py && sleep 0' 13731 1727203846.11152: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203846.11155: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203846.11157: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.11160: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.11209: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.11216: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.11255: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.26442: stdout chunk (state=3): >>> {"changed": true, "stdout": "00:00:5e:00:53:5d", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/ad_actor_system"], "start": "2024-09-24 14:50:46.260898", "end": "2024-09-24 14:50:46.263822", "delta": "0:00:00.002924", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/ad_actor_system", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203846.27850: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203846.27895: stderr chunk (state=3): >>><<< 13731 1727203846.27898: stdout chunk (state=3): >>><<< 13731 1727203846.27918: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "00:00:5e:00:53:5d", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/ad_actor_system"], "start": "2024-09-24 14:50:46.260898", "end": "2024-09-24 14:50:46.263822", "delta": "0:00:00.002924", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/ad_actor_system", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203846.27944: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/ad_actor_system', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203846.27948: _low_level_execute_command(): starting 13731 1727203846.27955: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203846.0200353-15163-197084158304019/ > /dev/null 2>&1 && sleep 0' 13731 1727203846.28540: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203846.28584: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.28587: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203846.28591: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.28596: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203846.28598: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203846.28663: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.28666: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.28669: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.28712: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.28754: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.30537: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.30584: stderr chunk (state=3): >>><<< 13731 1727203846.30588: stdout chunk (state=3): >>><<< 13731 1727203846.30613: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.30619: handler run complete 13731 1727203846.30635: Evaluated conditional (False): False 13731 1727203846.30826: variable 'bond_opt' from source: unknown 13731 1727203846.30834: variable 'result' from source: unknown 13731 1727203846.30845: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203846.30854: attempt loop complete, returning result 13731 1727203846.30869: variable 'bond_opt' from source: unknown 13731 1727203846.30916: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'ad_actor_system', 'value': '00:00:5e:00:53:5d'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "ad_actor_system", "value": "00:00:5e:00:53:5d" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/ad_actor_system" ], "delta": "0:00:00.002924", "end": "2024-09-24 14:50:46.263822", "rc": 0, "start": "2024-09-24 14:50:46.260898" } STDOUT: 00:00:5e:00:53:5d 13731 1727203846.31057: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203846.31065: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203846.31068: variable 'omit' from source: magic vars 13731 1727203846.31297: variable 'ansible_distribution_major_version' from source: facts 13731 1727203846.31300: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203846.31302: variable 'omit' from source: magic vars 13731 1727203846.31304: variable 'omit' from source: magic vars 13731 1727203846.31687: variable 'controller_device' from source: play vars 13731 1727203846.31690: variable 'bond_opt' from source: unknown 13731 1727203846.31693: variable 'omit' from source: magic vars 13731 1727203846.31705: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203846.31715: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203846.31725: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203846.31741: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203846.31800: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203846.31807: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203846.31887: Set connection var ansible_pipelining to False 13731 1727203846.31989: Set connection var ansible_shell_type to sh 13731 1727203846.32003: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203846.32023: Set connection var ansible_connection to ssh 13731 1727203846.32034: Set connection var ansible_shell_executable to /bin/sh 13731 1727203846.32045: Set connection var ansible_timeout to 10 13731 1727203846.32124: variable 'ansible_shell_executable' from source: unknown 13731 1727203846.32127: variable 'ansible_connection' from source: unknown 13731 1727203846.32129: variable 'ansible_module_compression' from source: unknown 13731 1727203846.32132: variable 'ansible_shell_type' from source: unknown 13731 1727203846.32134: variable 'ansible_shell_executable' from source: unknown 13731 1727203846.32136: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203846.32158: variable 'ansible_pipelining' from source: unknown 13731 1727203846.32163: variable 'ansible_timeout' from source: unknown 13731 1727203846.32165: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203846.32304: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203846.32308: variable 'omit' from source: magic vars 13731 1727203846.32310: starting attempt loop 13731 1727203846.32312: running the handler 13731 1727203846.32314: _low_level_execute_command(): starting 13731 1727203846.32316: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203846.32736: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.32749: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203846.32773: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.32820: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.32824: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.32830: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.32860: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.34414: stdout chunk (state=3): >>>/root <<< 13731 1727203846.34513: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.34537: stderr chunk (state=3): >>><<< 13731 1727203846.34540: stdout chunk (state=3): >>><<< 13731 1727203846.34554: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.34562: _low_level_execute_command(): starting 13731 1727203846.34569: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814 `" && echo ansible-tmp-1727203846.3455358-15163-16362810731814="` echo /root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814 `" ) && sleep 0' 13731 1727203846.35441: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.35457: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.37273: stdout chunk (state=3): >>>ansible-tmp-1727203846.3455358-15163-16362810731814=/root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814 <<< 13731 1727203846.37404: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.37408: stderr chunk (state=3): >>><<< 13731 1727203846.37410: stdout chunk (state=3): >>><<< 13731 1727203846.37602: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203846.3455358-15163-16362810731814=/root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.37606: variable 'ansible_module_compression' from source: unknown 13731 1727203846.37608: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203846.37611: variable 'ansible_facts' from source: unknown 13731 1727203846.37616: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/AnsiballZ_command.py 13731 1727203846.37865: Sending initial data 13731 1727203846.37877: Sent initial data (155 bytes) 13731 1727203846.38387: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203846.38400: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203846.38412: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.38427: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203846.38488: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.38528: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.38542: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.38560: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.38622: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.40172: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203846.40384: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/AnsiballZ_command.py" <<< 13731 1727203846.40387: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpfi0eah8a /root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/AnsiballZ_command.py <<< 13731 1727203846.40390: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpfi0eah8a" to remote "/root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/AnsiballZ_command.py" <<< 13731 1727203846.41169: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.41223: stderr chunk (state=3): >>><<< 13731 1727203846.41233: stdout chunk (state=3): >>><<< 13731 1727203846.41280: done transferring module to remote 13731 1727203846.41294: _low_level_execute_command(): starting 13731 1727203846.41365: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/ /root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/AnsiballZ_command.py && sleep 0' 13731 1727203846.41969: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.41973: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203846.41979: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.41990: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.42048: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.42077: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.42141: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.43835: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.43863: stderr chunk (state=3): >>><<< 13731 1727203846.43880: stdout chunk (state=3): >>><<< 13731 1727203846.44195: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.44198: _low_level_execute_command(): starting 13731 1727203846.44200: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/AnsiballZ_command.py && sleep 0' 13731 1727203846.45200: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203846.45213: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203846.45224: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.45291: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.45304: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.45500: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.45563: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.60717: stdout chunk (state=3): >>> {"changed": true, "stdout": "stable 0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/ad_select"], "start": "2024-09-24 14:50:46.603396", "end": "2024-09-24 14:50:46.606460", "delta": "0:00:00.003064", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/ad_select", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203846.62130: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203846.62157: stderr chunk (state=3): >>><<< 13731 1727203846.62167: stdout chunk (state=3): >>><<< 13731 1727203846.62194: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "stable 0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/ad_select"], "start": "2024-09-24 14:50:46.603396", "end": "2024-09-24 14:50:46.606460", "delta": "0:00:00.003064", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/ad_select", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203846.62239: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/ad_select', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203846.62344: _low_level_execute_command(): starting 13731 1727203846.62420: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203846.3455358-15163-16362810731814/ > /dev/null 2>&1 && sleep 0' 13731 1727203846.63472: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203846.63659: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.63825: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.63855: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.66182: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.66188: stdout chunk (state=3): >>><<< 13731 1727203846.66190: stderr chunk (state=3): >>><<< 13731 1727203846.66193: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.66195: handler run complete 13731 1727203846.66197: Evaluated conditional (False): False 13731 1727203846.66641: variable 'bond_opt' from source: unknown 13731 1727203846.66644: variable 'result' from source: unknown 13731 1727203846.66647: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203846.66649: attempt loop complete, returning result 13731 1727203846.66651: variable 'bond_opt' from source: unknown 13731 1727203846.66765: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'ad_select', 'value': 'stable'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "ad_select", "value": "stable" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/ad_select" ], "delta": "0:00:00.003064", "end": "2024-09-24 14:50:46.606460", "rc": 0, "start": "2024-09-24 14:50:46.603396" } STDOUT: stable 0 13731 1727203846.67281: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203846.67284: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203846.67287: variable 'omit' from source: magic vars 13731 1727203846.67514: variable 'ansible_distribution_major_version' from source: facts 13731 1727203846.67527: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203846.67536: variable 'omit' from source: magic vars 13731 1727203846.67556: variable 'omit' from source: magic vars 13731 1727203846.68081: variable 'controller_device' from source: play vars 13731 1727203846.68085: variable 'bond_opt' from source: unknown 13731 1727203846.68087: variable 'omit' from source: magic vars 13731 1727203846.68089: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203846.68091: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203846.68094: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203846.68096: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203846.68098: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203846.68100: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203846.68228: Set connection var ansible_pipelining to False 13731 1727203846.68482: Set connection var ansible_shell_type to sh 13731 1727203846.68487: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203846.68489: Set connection var ansible_connection to ssh 13731 1727203846.68492: Set connection var ansible_shell_executable to /bin/sh 13731 1727203846.68494: Set connection var ansible_timeout to 10 13731 1727203846.68496: variable 'ansible_shell_executable' from source: unknown 13731 1727203846.68498: variable 'ansible_connection' from source: unknown 13731 1727203846.68500: variable 'ansible_module_compression' from source: unknown 13731 1727203846.68502: variable 'ansible_shell_type' from source: unknown 13731 1727203846.68504: variable 'ansible_shell_executable' from source: unknown 13731 1727203846.68506: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203846.68508: variable 'ansible_pipelining' from source: unknown 13731 1727203846.68510: variable 'ansible_timeout' from source: unknown 13731 1727203846.68512: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203846.68654: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203846.68709: variable 'omit' from source: magic vars 13731 1727203846.68719: starting attempt loop 13731 1727203846.68726: running the handler 13731 1727203846.68790: _low_level_execute_command(): starting 13731 1727203846.68810: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203846.69877: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.69937: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203846.70114: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.70169: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.70221: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.70336: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.70386: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.72029: stdout chunk (state=3): >>>/root <<< 13731 1727203846.72118: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.72163: stderr chunk (state=3): >>><<< 13731 1727203846.72166: stdout chunk (state=3): >>><<< 13731 1727203846.72382: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.72386: _low_level_execute_command(): starting 13731 1727203846.72390: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561 `" && echo ansible-tmp-1727203846.7218235-15163-59186014645561="` echo /root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561 `" ) && sleep 0' 13731 1727203846.73215: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203846.73245: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203846.73610: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.73703: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.75562: stdout chunk (state=3): >>>ansible-tmp-1727203846.7218235-15163-59186014645561=/root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561 <<< 13731 1727203846.75673: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.75886: stderr chunk (state=3): >>><<< 13731 1727203846.75889: stdout chunk (state=3): >>><<< 13731 1727203846.75892: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203846.7218235-15163-59186014645561=/root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.75894: variable 'ansible_module_compression' from source: unknown 13731 1727203846.75896: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203846.75898: variable 'ansible_facts' from source: unknown 13731 1727203846.76056: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/AnsiballZ_command.py 13731 1727203846.76335: Sending initial data 13731 1727203846.76351: Sent initial data (155 bytes) 13731 1727203846.77763: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203846.77783: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.77809: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.77879: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.79436: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" <<< 13731 1727203846.79460: stderr chunk (state=3): >>>debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203846.79601: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/AnsiballZ_command.py" <<< 13731 1727203846.79605: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp_49p2pi0 /root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/AnsiballZ_command.py <<< 13731 1727203846.79778: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp_49p2pi0" to remote "/root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/AnsiballZ_command.py" <<< 13731 1727203846.79793: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/AnsiballZ_command.py" <<< 13731 1727203846.81106: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.81319: stderr chunk (state=3): >>><<< 13731 1727203846.81322: stdout chunk (state=3): >>><<< 13731 1727203846.81324: done transferring module to remote 13731 1727203846.81327: _low_level_execute_command(): starting 13731 1727203846.81329: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/ /root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/AnsiballZ_command.py && sleep 0' 13731 1727203846.82345: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203846.82350: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203846.82352: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203846.82569: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.82678: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.82734: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203846.84445: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203846.84480: stderr chunk (state=3): >>><<< 13731 1727203846.84511: stdout chunk (state=3): >>><<< 13731 1727203846.84783: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203846.84786: _low_level_execute_command(): starting 13731 1727203846.84788: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/AnsiballZ_command.py && sleep 0' 13731 1727203846.86497: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203846.86505: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203846.86588: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.01856: stdout chunk (state=3): >>> {"changed": true, "stdout": "1023", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/ad_user_port_key"], "start": "2024-09-24 14:50:47.015121", "end": "2024-09-24 14:50:47.017993", "delta": "0:00:00.002872", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/ad_user_port_key", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203847.03373: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203847.03381: stdout chunk (state=3): >>><<< 13731 1727203847.03386: stderr chunk (state=3): >>><<< 13731 1727203847.03404: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "1023", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/ad_user_port_key"], "start": "2024-09-24 14:50:47.015121", "end": "2024-09-24 14:50:47.017993", "delta": "0:00:00.002872", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/ad_user_port_key", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203847.03432: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/ad_user_port_key', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203847.03437: _low_level_execute_command(): starting 13731 1727203847.03442: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203846.7218235-15163-59186014645561/ > /dev/null 2>&1 && sleep 0' 13731 1727203847.04024: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.04033: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.04044: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.04058: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.04074: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.04121: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203847.04128: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.04131: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203847.04134: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203847.04136: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203847.04138: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.04140: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.04142: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.04152: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.04228: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203847.04232: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.04234: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.04249: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.04274: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.04310: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.06148: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.06152: stdout chunk (state=3): >>><<< 13731 1727203847.06154: stderr chunk (state=3): >>><<< 13731 1727203847.06380: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.06384: handler run complete 13731 1727203847.06386: Evaluated conditional (False): False 13731 1727203847.06388: variable 'bond_opt' from source: unknown 13731 1727203847.06390: variable 'result' from source: unknown 13731 1727203847.06391: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203847.06393: attempt loop complete, returning result 13731 1727203847.06413: variable 'bond_opt' from source: unknown 13731 1727203847.06475: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'ad_user_port_key', 'value': '1023'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "ad_user_port_key", "value": "1023" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/ad_user_port_key" ], "delta": "0:00:00.002872", "end": "2024-09-24 14:50:47.017993", "rc": 0, "start": "2024-09-24 14:50:47.015121" } STDOUT: 1023 13731 1727203847.06731: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.06745: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.06815: variable 'omit' from source: magic vars 13731 1727203847.06934: variable 'ansible_distribution_major_version' from source: facts 13731 1727203847.06955: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203847.06967: variable 'omit' from source: magic vars 13731 1727203847.06989: variable 'omit' from source: magic vars 13731 1727203847.07184: variable 'controller_device' from source: play vars 13731 1727203847.07193: variable 'bond_opt' from source: unknown 13731 1727203847.07215: variable 'omit' from source: magic vars 13731 1727203847.07239: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203847.07257: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203847.07285: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203847.07359: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203847.07366: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.07368: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.07413: Set connection var ansible_pipelining to False 13731 1727203847.07423: Set connection var ansible_shell_type to sh 13731 1727203847.07433: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203847.07443: Set connection var ansible_connection to ssh 13731 1727203847.07452: Set connection var ansible_shell_executable to /bin/sh 13731 1727203847.07470: Set connection var ansible_timeout to 10 13731 1727203847.07505: variable 'ansible_shell_executable' from source: unknown 13731 1727203847.07512: variable 'ansible_connection' from source: unknown 13731 1727203847.07519: variable 'ansible_module_compression' from source: unknown 13731 1727203847.07525: variable 'ansible_shell_type' from source: unknown 13731 1727203847.07579: variable 'ansible_shell_executable' from source: unknown 13731 1727203847.07590: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.07593: variable 'ansible_pipelining' from source: unknown 13731 1727203847.07882: variable 'ansible_timeout' from source: unknown 13731 1727203847.07885: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.07887: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203847.07889: variable 'omit' from source: magic vars 13731 1727203847.07891: starting attempt loop 13731 1727203847.07893: running the handler 13731 1727203847.07894: _low_level_execute_command(): starting 13731 1727203847.07896: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203847.08519: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.08571: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.08637: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.08684: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.08705: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.08777: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.10377: stdout chunk (state=3): >>>/root <<< 13731 1727203847.10522: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.10526: stdout chunk (state=3): >>><<< 13731 1727203847.10528: stderr chunk (state=3): >>><<< 13731 1727203847.10622: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.10630: _low_level_execute_command(): starting 13731 1727203847.10632: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736 `" && echo ansible-tmp-1727203847.1054454-15163-50279624763736="` echo /root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736 `" ) && sleep 0' 13731 1727203847.11207: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.11231: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.11248: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.11268: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.11289: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.11301: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203847.11328: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.11348: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203847.11390: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.11463: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.11486: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.11506: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.11571: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.13452: stdout chunk (state=3): >>>ansible-tmp-1727203847.1054454-15163-50279624763736=/root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736 <<< 13731 1727203847.13610: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.13614: stdout chunk (state=3): >>><<< 13731 1727203847.13617: stderr chunk (state=3): >>><<< 13731 1727203847.13781: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203847.1054454-15163-50279624763736=/root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.13785: variable 'ansible_module_compression' from source: unknown 13731 1727203847.13792: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203847.13794: variable 'ansible_facts' from source: unknown 13731 1727203847.13806: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/AnsiballZ_command.py 13731 1727203847.14020: Sending initial data 13731 1727203847.14029: Sent initial data (155 bytes) 13731 1727203847.14545: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.14562: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.14577: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.14593: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.14684: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.14699: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.14757: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.16257: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203847.16309: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203847.16365: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpw32_2bcd /root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/AnsiballZ_command.py <<< 13731 1727203847.16393: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/AnsiballZ_command.py" debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpw32_2bcd" to remote "/root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/AnsiballZ_command.py" <<< 13731 1727203847.17156: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.17190: stdout chunk (state=3): >>><<< 13731 1727203847.17217: stderr chunk (state=3): >>><<< 13731 1727203847.17291: done transferring module to remote 13731 1727203847.17373: _low_level_execute_command(): starting 13731 1727203847.17388: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/ /root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/AnsiballZ_command.py && sleep 0' 13731 1727203847.18058: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.18095: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.18118: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.18136: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.18212: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.20049: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.20059: stdout chunk (state=3): >>><<< 13731 1727203847.20070: stderr chunk (state=3): >>><<< 13731 1727203847.20130: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.20139: _low_level_execute_command(): starting 13731 1727203847.20147: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/AnsiballZ_command.py && sleep 0' 13731 1727203847.20716: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.20733: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.20749: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.20766: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.20788: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.20800: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203847.20815: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.20888: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.20922: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.20947: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.20964: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.21038: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.36279: stdout chunk (state=3): >>> {"changed": true, "stdout": "1", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/all_slaves_active"], "start": "2024-09-24 14:50:47.359116", "end": "2024-09-24 14:50:47.362060", "delta": "0:00:00.002944", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/all_slaves_active", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203847.37883: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203847.37887: stderr chunk (state=3): >>><<< 13731 1727203847.37889: stdout chunk (state=3): >>><<< 13731 1727203847.37892: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "1", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/all_slaves_active"], "start": "2024-09-24 14:50:47.359116", "end": "2024-09-24 14:50:47.362060", "delta": "0:00:00.002944", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/all_slaves_active", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203847.37894: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/all_slaves_active', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203847.37896: _low_level_execute_command(): starting 13731 1727203847.37898: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203847.1054454-15163-50279624763736/ > /dev/null 2>&1 && sleep 0' 13731 1727203847.38535: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.38591: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.38598: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203847.38608: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.38696: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.38719: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.38735: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.38798: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.40658: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.40661: stdout chunk (state=3): >>><<< 13731 1727203847.40664: stderr chunk (state=3): >>><<< 13731 1727203847.40881: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.40884: handler run complete 13731 1727203847.40886: Evaluated conditional (False): False 13731 1727203847.40889: variable 'bond_opt' from source: unknown 13731 1727203847.40890: variable 'result' from source: unknown 13731 1727203847.40892: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203847.40908: attempt loop complete, returning result 13731 1727203847.40930: variable 'bond_opt' from source: unknown 13731 1727203847.40997: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'all_slaves_active', 'value': '1'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "all_slaves_active", "value": "1" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/all_slaves_active" ], "delta": "0:00:00.002944", "end": "2024-09-24 14:50:47.362060", "rc": 0, "start": "2024-09-24 14:50:47.359116" } STDOUT: 1 13731 1727203847.41213: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.41226: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.41239: variable 'omit' from source: magic vars 13731 1727203847.41391: variable 'ansible_distribution_major_version' from source: facts 13731 1727203847.41401: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203847.41409: variable 'omit' from source: magic vars 13731 1727203847.41424: variable 'omit' from source: magic vars 13731 1727203847.41780: variable 'controller_device' from source: play vars 13731 1727203847.41783: variable 'bond_opt' from source: unknown 13731 1727203847.41786: variable 'omit' from source: magic vars 13731 1727203847.41788: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203847.41790: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203847.41792: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203847.41794: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203847.41796: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.41799: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.41801: Set connection var ansible_pipelining to False 13731 1727203847.41802: Set connection var ansible_shell_type to sh 13731 1727203847.41804: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203847.41806: Set connection var ansible_connection to ssh 13731 1727203847.41808: Set connection var ansible_shell_executable to /bin/sh 13731 1727203847.41810: Set connection var ansible_timeout to 10 13731 1727203847.41812: variable 'ansible_shell_executable' from source: unknown 13731 1727203847.41814: variable 'ansible_connection' from source: unknown 13731 1727203847.41823: variable 'ansible_module_compression' from source: unknown 13731 1727203847.41830: variable 'ansible_shell_type' from source: unknown 13731 1727203847.41837: variable 'ansible_shell_executable' from source: unknown 13731 1727203847.41843: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.41851: variable 'ansible_pipelining' from source: unknown 13731 1727203847.41858: variable 'ansible_timeout' from source: unknown 13731 1727203847.41866: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.41965: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203847.41982: variable 'omit' from source: magic vars 13731 1727203847.41991: starting attempt loop 13731 1727203847.41998: running the handler 13731 1727203847.42009: _low_level_execute_command(): starting 13731 1727203847.42016: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203847.42642: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.42658: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.42677: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.42704: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.42804: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.42817: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.42891: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.44437: stdout chunk (state=3): >>>/root <<< 13731 1727203847.44603: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.44607: stdout chunk (state=3): >>><<< 13731 1727203847.44609: stderr chunk (state=3): >>><<< 13731 1727203847.44709: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.44713: _low_level_execute_command(): starting 13731 1727203847.44716: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692 `" && echo ansible-tmp-1727203847.4462504-15163-23787549835692="` echo /root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692 `" ) && sleep 0' 13731 1727203847.45397: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.45405: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.45493: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.45560: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.47380: stdout chunk (state=3): >>>ansible-tmp-1727203847.4462504-15163-23787549835692=/root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692 <<< 13731 1727203847.47686: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.47690: stdout chunk (state=3): >>><<< 13731 1727203847.47692: stderr chunk (state=3): >>><<< 13731 1727203847.47695: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203847.4462504-15163-23787549835692=/root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.47697: variable 'ansible_module_compression' from source: unknown 13731 1727203847.47699: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203847.47701: variable 'ansible_facts' from source: unknown 13731 1727203847.47728: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/AnsiballZ_command.py 13731 1727203847.47902: Sending initial data 13731 1727203847.47906: Sent initial data (155 bytes) 13731 1727203847.48513: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.48526: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.48595: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.48642: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.48665: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.48715: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.48747: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.50274: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203847.50318: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203847.50360: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp7ea4vt5b /root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/AnsiballZ_command.py <<< 13731 1727203847.50382: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/AnsiballZ_command.py" <<< 13731 1727203847.50414: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp7ea4vt5b" to remote "/root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/AnsiballZ_command.py" <<< 13731 1727203847.51229: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.51232: stdout chunk (state=3): >>><<< 13731 1727203847.51234: stderr chunk (state=3): >>><<< 13731 1727203847.51242: done transferring module to remote 13731 1727203847.51252: _low_level_execute_command(): starting 13731 1727203847.51257: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/ /root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/AnsiballZ_command.py && sleep 0' 13731 1727203847.51885: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.51893: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.51904: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.51992: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.51997: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.52000: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203847.52002: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.52004: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203847.52006: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203847.52009: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203847.52011: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.52013: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.52015: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.52017: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.52022: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203847.52036: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.52119: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.52132: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.52197: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.54377: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.54381: stdout chunk (state=3): >>><<< 13731 1727203847.54384: stderr chunk (state=3): >>><<< 13731 1727203847.54387: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.54389: _low_level_execute_command(): starting 13731 1727203847.54391: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/AnsiballZ_command.py && sleep 0' 13731 1727203847.54887: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.54890: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.54892: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.54895: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.54897: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.54899: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203847.54901: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.54903: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203847.54917: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203847.54920: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203847.54926: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.54935: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.54946: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.54954: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.54960: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203847.54973: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.55040: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.55050: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.55067: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.55138: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.70584: stdout chunk (state=3): >>> {"changed": true, "stdout": "0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/downdelay"], "start": "2024-09-24 14:50:47.699182", "end": "2024-09-24 14:50:47.702251", "delta": "0:00:00.003069", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/downdelay", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203847.71953: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.71996: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203847.72280: stderr chunk (state=3): >>><<< 13731 1727203847.72284: stdout chunk (state=3): >>><<< 13731 1727203847.72287: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/downdelay"], "start": "2024-09-24 14:50:47.699182", "end": "2024-09-24 14:50:47.702251", "delta": "0:00:00.003069", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/downdelay", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203847.72289: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/downdelay', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203847.72291: _low_level_execute_command(): starting 13731 1727203847.72293: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203847.4462504-15163-23787549835692/ > /dev/null 2>&1 && sleep 0' 13731 1727203847.73643: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.73827: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.73943: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.73990: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.75780: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.75969: stderr chunk (state=3): >>><<< 13731 1727203847.75973: stdout chunk (state=3): >>><<< 13731 1727203847.75978: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.75981: handler run complete 13731 1727203847.75984: Evaluated conditional (False): False 13731 1727203847.76329: variable 'bond_opt' from source: unknown 13731 1727203847.76338: variable 'result' from source: unknown 13731 1727203847.76364: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203847.76367: attempt loop complete, returning result 13731 1727203847.76385: variable 'bond_opt' from source: unknown 13731 1727203847.76448: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'downdelay', 'value': '0'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "downdelay", "value": "0" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/downdelay" ], "delta": "0:00:00.003069", "end": "2024-09-24 14:50:47.702251", "rc": 0, "start": "2024-09-24 14:50:47.699182" } STDOUT: 0 13731 1727203847.77042: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.77048: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.77059: variable 'omit' from source: magic vars 13731 1727203847.77754: variable 'ansible_distribution_major_version' from source: facts 13731 1727203847.77759: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203847.77765: variable 'omit' from source: magic vars 13731 1727203847.77883: variable 'omit' from source: magic vars 13731 1727203847.78341: variable 'controller_device' from source: play vars 13731 1727203847.78558: variable 'bond_opt' from source: unknown 13731 1727203847.78564: variable 'omit' from source: magic vars 13731 1727203847.78567: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203847.78569: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203847.78571: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203847.78573: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203847.78577: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.78579: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.78982: Set connection var ansible_pipelining to False 13731 1727203847.78985: Set connection var ansible_shell_type to sh 13731 1727203847.78988: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203847.78990: Set connection var ansible_connection to ssh 13731 1727203847.78992: Set connection var ansible_shell_executable to /bin/sh 13731 1727203847.78994: Set connection var ansible_timeout to 10 13731 1727203847.78996: variable 'ansible_shell_executable' from source: unknown 13731 1727203847.78997: variable 'ansible_connection' from source: unknown 13731 1727203847.78999: variable 'ansible_module_compression' from source: unknown 13731 1727203847.79001: variable 'ansible_shell_type' from source: unknown 13731 1727203847.79003: variable 'ansible_shell_executable' from source: unknown 13731 1727203847.79005: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203847.79007: variable 'ansible_pipelining' from source: unknown 13731 1727203847.79009: variable 'ansible_timeout' from source: unknown 13731 1727203847.79011: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203847.79119: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203847.79126: variable 'omit' from source: magic vars 13731 1727203847.79129: starting attempt loop 13731 1727203847.79131: running the handler 13731 1727203847.79139: _low_level_execute_command(): starting 13731 1727203847.79141: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203847.80372: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.80389: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.80510: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.80516: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.80618: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.80791: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.80892: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.82442: stdout chunk (state=3): >>>/root <<< 13731 1727203847.82667: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.82670: stdout chunk (state=3): >>><<< 13731 1727203847.82672: stderr chunk (state=3): >>><<< 13731 1727203847.82691: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.82699: _low_level_execute_command(): starting 13731 1727203847.82769: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578 `" && echo ansible-tmp-1727203847.826898-15163-160555065005578="` echo /root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578 `" ) && sleep 0' 13731 1727203847.83959: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.83965: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.83968: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.83971: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.83973: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203847.83975: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203847.83979: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.83981: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203847.83983: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203847.83985: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203847.83987: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.83989: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.84046: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.84056: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.84072: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.84186: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.85968: stdout chunk (state=3): >>>ansible-tmp-1727203847.826898-15163-160555065005578=/root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578 <<< 13731 1727203847.86181: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.86185: stdout chunk (state=3): >>><<< 13731 1727203847.86192: stderr chunk (state=3): >>><<< 13731 1727203847.86212: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203847.826898-15163-160555065005578=/root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.86294: variable 'ansible_module_compression' from source: unknown 13731 1727203847.86329: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203847.86349: variable 'ansible_facts' from source: unknown 13731 1727203847.86536: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/AnsiballZ_command.py 13731 1727203847.86886: Sending initial data 13731 1727203847.86889: Sent initial data (155 bytes) 13731 1727203847.88335: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203847.88351: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203847.88357: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.88442: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.88599: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.88664: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.88668: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.90172: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203847.90218: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203847.90364: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpl0e1pon9 /root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/AnsiballZ_command.py <<< 13731 1727203847.90368: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/AnsiballZ_command.py" <<< 13731 1727203847.90370: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpl0e1pon9" to remote "/root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/AnsiballZ_command.py" <<< 13731 1727203847.91955: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.91983: stderr chunk (state=3): >>><<< 13731 1727203847.91986: stdout chunk (state=3): >>><<< 13731 1727203847.92047: done transferring module to remote 13731 1727203847.92082: _low_level_execute_command(): starting 13731 1727203847.92176: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/ /root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/AnsiballZ_command.py && sleep 0' 13731 1727203847.92986: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203847.93082: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203847.93097: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.93158: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203847.93181: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.93208: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.93303: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203847.95384: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203847.95388: stdout chunk (state=3): >>><<< 13731 1727203847.95391: stderr chunk (state=3): >>><<< 13731 1727203847.95393: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203847.95396: _low_level_execute_command(): starting 13731 1727203847.95398: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/AnsiballZ_command.py && sleep 0' 13731 1727203847.95850: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203847.95889: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203847.95982: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203847.96068: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203847.96129: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.11484: stdout chunk (state=3): >>> {"changed": true, "stdout": "slow 0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/lacp_rate"], "start": "2024-09-24 14:50:48.111408", "end": "2024-09-24 14:50:48.114219", "delta": "0:00:00.002811", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/lacp_rate", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203848.12856: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.12911: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203848.13019: stderr chunk (state=3): >>><<< 13731 1727203848.13022: stdout chunk (state=3): >>><<< 13731 1727203848.13043: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "slow 0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/lacp_rate"], "start": "2024-09-24 14:50:48.111408", "end": "2024-09-24 14:50:48.114219", "delta": "0:00:00.002811", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/lacp_rate", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203848.13072: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/lacp_rate', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203848.13078: _low_level_execute_command(): starting 13731 1727203848.13084: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203847.826898-15163-160555065005578/ > /dev/null 2>&1 && sleep 0' 13731 1727203848.14320: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.14493: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.14554: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.14693: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.16524: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.16527: stdout chunk (state=3): >>><<< 13731 1727203848.16533: stderr chunk (state=3): >>><<< 13731 1727203848.16553: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.16556: handler run complete 13731 1727203848.16687: Evaluated conditional (False): False 13731 1727203848.17130: variable 'bond_opt' from source: unknown 13731 1727203848.17136: variable 'result' from source: unknown 13731 1727203848.17209: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203848.17212: attempt loop complete, returning result 13731 1727203848.17214: variable 'bond_opt' from source: unknown 13731 1727203848.17472: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'lacp_rate', 'value': 'slow'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "lacp_rate", "value": "slow" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/lacp_rate" ], "delta": "0:00:00.002811", "end": "2024-09-24 14:50:48.114219", "rc": 0, "start": "2024-09-24 14:50:48.111408" } STDOUT: slow 0 13731 1727203848.17778: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.17781: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.17992: variable 'omit' from source: magic vars 13731 1727203848.18229: variable 'ansible_distribution_major_version' from source: facts 13731 1727203848.18235: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203848.18238: variable 'omit' from source: magic vars 13731 1727203848.18267: variable 'omit' from source: magic vars 13731 1727203848.18634: variable 'controller_device' from source: play vars 13731 1727203848.18637: variable 'bond_opt' from source: unknown 13731 1727203848.18659: variable 'omit' from source: magic vars 13731 1727203848.18729: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203848.18739: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203848.18746: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203848.18765: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203848.18768: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.18771: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.18991: Set connection var ansible_pipelining to False 13731 1727203848.18994: Set connection var ansible_shell_type to sh 13731 1727203848.18996: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203848.18998: Set connection var ansible_connection to ssh 13731 1727203848.19000: Set connection var ansible_shell_executable to /bin/sh 13731 1727203848.19002: Set connection var ansible_timeout to 10 13731 1727203848.19082: variable 'ansible_shell_executable' from source: unknown 13731 1727203848.19085: variable 'ansible_connection' from source: unknown 13731 1727203848.19087: variable 'ansible_module_compression' from source: unknown 13731 1727203848.19089: variable 'ansible_shell_type' from source: unknown 13731 1727203848.19091: variable 'ansible_shell_executable' from source: unknown 13731 1727203848.19094: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.19095: variable 'ansible_pipelining' from source: unknown 13731 1727203848.19097: variable 'ansible_timeout' from source: unknown 13731 1727203848.19099: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.19242: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203848.19250: variable 'omit' from source: magic vars 13731 1727203848.19253: starting attempt loop 13731 1727203848.19256: running the handler 13731 1727203848.19265: _low_level_execute_command(): starting 13731 1727203848.19407: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203848.20955: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203848.20959: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.20965: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203848.20967: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.21088: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.21131: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.22696: stdout chunk (state=3): >>>/root <<< 13731 1727203848.23174: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.23179: stderr chunk (state=3): >>><<< 13731 1727203848.23182: stdout chunk (state=3): >>><<< 13731 1727203848.23184: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.23186: _low_level_execute_command(): starting 13731 1727203848.23188: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780 `" && echo ansible-tmp-1727203848.2313066-15163-10940300590780="` echo /root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780 `" ) && sleep 0' 13731 1727203848.24394: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203848.24397: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203848.24400: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203848.24471: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203848.24478: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203848.24592: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203848.24637: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.24645: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.24765: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.26616: stdout chunk (state=3): >>>ansible-tmp-1727203848.2313066-15163-10940300590780=/root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780 <<< 13731 1727203848.26773: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.26927: stderr chunk (state=3): >>><<< 13731 1727203848.26930: stdout chunk (state=3): >>><<< 13731 1727203848.26953: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203848.2313066-15163-10940300590780=/root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.26972: variable 'ansible_module_compression' from source: unknown 13731 1727203848.27012: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203848.27283: variable 'ansible_facts' from source: unknown 13731 1727203848.27291: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/AnsiballZ_command.py 13731 1727203848.27421: Sending initial data 13731 1727203848.27424: Sent initial data (155 bytes) 13731 1727203848.28892: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.28993: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203848.28997: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.29000: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.29066: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.30575: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203848.30634: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203848.30713: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpiqg77qsp /root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/AnsiballZ_command.py <<< 13731 1727203848.30717: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/AnsiballZ_command.py" <<< 13731 1727203848.31089: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpiqg77qsp" to remote "/root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/AnsiballZ_command.py" <<< 13731 1727203848.32582: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.32586: stderr chunk (state=3): >>><<< 13731 1727203848.32588: stdout chunk (state=3): >>><<< 13731 1727203848.32590: done transferring module to remote 13731 1727203848.32592: _low_level_execute_command(): starting 13731 1727203848.32594: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/ /root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/AnsiballZ_command.py && sleep 0' 13731 1727203848.34063: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203848.34158: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.34248: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203848.34267: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.34318: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.34355: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.36292: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.36296: stdout chunk (state=3): >>><<< 13731 1727203848.36299: stderr chunk (state=3): >>><<< 13731 1727203848.36435: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.36438: _low_level_execute_command(): starting 13731 1727203848.36440: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/AnsiballZ_command.py && sleep 0' 13731 1727203848.37662: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203848.37825: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.37936: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.38010: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.53304: stdout chunk (state=3): >>> {"changed": true, "stdout": "128", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/lp_interval"], "start": "2024-09-24 14:50:48.529467", "end": "2024-09-24 14:50:48.532484", "delta": "0:00:00.003017", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/lp_interval", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203848.54793: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.54855: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203848.54912: stdout chunk (state=3): >>><<< 13731 1727203848.54920: stderr chunk (state=3): >>><<< 13731 1727203848.55287: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "128", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/lp_interval"], "start": "2024-09-24 14:50:48.529467", "end": "2024-09-24 14:50:48.532484", "delta": "0:00:00.003017", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/lp_interval", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203848.55290: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/lp_interval', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203848.55293: _low_level_execute_command(): starting 13731 1727203848.55295: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203848.2313066-15163-10940300590780/ > /dev/null 2>&1 && sleep 0' 13731 1727203848.55862: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203848.55879: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203848.55896: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203848.55992: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.56011: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203848.56026: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.56268: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.56321: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.58181: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.58199: stdout chunk (state=3): >>><<< 13731 1727203848.58211: stderr chunk (state=3): >>><<< 13731 1727203848.58230: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.58239: handler run complete 13731 1727203848.58265: Evaluated conditional (False): False 13731 1727203848.58428: variable 'bond_opt' from source: unknown 13731 1727203848.58439: variable 'result' from source: unknown 13731 1727203848.58456: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203848.58472: attempt loop complete, returning result 13731 1727203848.58498: variable 'bond_opt' from source: unknown 13731 1727203848.58575: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'lp_interval', 'value': '128'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "lp_interval", "value": "128" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/lp_interval" ], "delta": "0:00:00.003017", "end": "2024-09-24 14:50:48.532484", "rc": 0, "start": "2024-09-24 14:50:48.529467" } STDOUT: 128 13731 1727203848.59065: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.59068: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.59071: variable 'omit' from source: magic vars 13731 1727203848.59701: variable 'ansible_distribution_major_version' from source: facts 13731 1727203848.59706: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203848.59716: variable 'omit' from source: magic vars 13731 1727203848.59920: variable 'omit' from source: magic vars 13731 1727203848.60519: variable 'controller_device' from source: play vars 13731 1727203848.60522: variable 'bond_opt' from source: unknown 13731 1727203848.60525: variable 'omit' from source: magic vars 13731 1727203848.60527: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203848.60529: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203848.60531: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203848.60538: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203848.60540: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.60542: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.60749: Set connection var ansible_pipelining to False 13731 1727203848.60760: Set connection var ansible_shell_type to sh 13731 1727203848.60770: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203848.60783: Set connection var ansible_connection to ssh 13731 1727203848.60793: Set connection var ansible_shell_executable to /bin/sh 13731 1727203848.60801: Set connection var ansible_timeout to 10 13731 1727203848.60824: variable 'ansible_shell_executable' from source: unknown 13731 1727203848.60851: variable 'ansible_connection' from source: unknown 13731 1727203848.60887: variable 'ansible_module_compression' from source: unknown 13731 1727203848.60895: variable 'ansible_shell_type' from source: unknown 13731 1727203848.60903: variable 'ansible_shell_executable' from source: unknown 13731 1727203848.60910: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.60919: variable 'ansible_pipelining' from source: unknown 13731 1727203848.60960: variable 'ansible_timeout' from source: unknown 13731 1727203848.60969: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.61184: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203848.61192: variable 'omit' from source: magic vars 13731 1727203848.61200: starting attempt loop 13731 1727203848.61207: running the handler 13731 1727203848.61218: _low_level_execute_command(): starting 13731 1727203848.61226: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203848.62452: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203848.62483: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.62592: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.62770: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.62826: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.64463: stdout chunk (state=3): >>>/root <<< 13731 1727203848.64581: stdout chunk (state=3): >>><<< 13731 1727203848.64585: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.64587: stderr chunk (state=3): >>><<< 13731 1727203848.64590: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.64602: _low_level_execute_command(): starting 13731 1727203848.64612: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190 `" && echo ansible-tmp-1727203848.6459277-15163-142166745232190="` echo /root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190 `" ) && sleep 0' 13731 1727203848.65932: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203848.65993: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.66174: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.66189: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.66540: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.68395: stdout chunk (state=3): >>>ansible-tmp-1727203848.6459277-15163-142166745232190=/root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190 <<< 13731 1727203848.68549: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.68609: stderr chunk (state=3): >>><<< 13731 1727203848.68625: stdout chunk (state=3): >>><<< 13731 1727203848.68783: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203848.6459277-15163-142166745232190=/root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.68791: variable 'ansible_module_compression' from source: unknown 13731 1727203848.68794: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203848.68796: variable 'ansible_facts' from source: unknown 13731 1727203848.68817: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/AnsiballZ_command.py 13731 1727203848.69020: Sending initial data 13731 1727203848.69031: Sent initial data (156 bytes) 13731 1727203848.69712: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203848.69772: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.69836: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203848.69856: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.69877: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.69991: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.71459: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203848.71496: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203848.71642: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/AnsiballZ_command.py" <<< 13731 1727203848.71645: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmptx_nvrfw /root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/AnsiballZ_command.py <<< 13731 1727203848.71660: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmptx_nvrfw" to remote "/root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/AnsiballZ_command.py" <<< 13731 1727203848.72651: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.72695: stderr chunk (state=3): >>><<< 13731 1727203848.72733: stdout chunk (state=3): >>><<< 13731 1727203848.72830: done transferring module to remote 13731 1727203848.73059: _low_level_execute_command(): starting 13731 1727203848.73062: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/ /root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/AnsiballZ_command.py && sleep 0' 13731 1727203848.73627: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203848.73630: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203848.73632: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.73634: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203848.73636: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.73700: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.73718: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.73782: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.75581: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.75585: stderr chunk (state=3): >>><<< 13731 1727203848.75587: stdout chunk (state=3): >>><<< 13731 1727203848.75590: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.75593: _low_level_execute_command(): starting 13731 1727203848.75595: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/AnsiballZ_command.py && sleep 0' 13731 1727203848.76190: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203848.76198: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203848.76243: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.76306: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203848.76316: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.76338: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.76404: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.91664: stdout chunk (state=3): >>> {"changed": true, "stdout": "110", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/miimon"], "start": "2024-09-24 14:50:48.913000", "end": "2024-09-24 14:50:48.916006", "delta": "0:00:00.003006", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/miimon", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203848.93274: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203848.93393: stdout chunk (state=3): >>><<< 13731 1727203848.93396: stderr chunk (state=3): >>><<< 13731 1727203848.93398: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "110", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/miimon"], "start": "2024-09-24 14:50:48.913000", "end": "2024-09-24 14:50:48.916006", "delta": "0:00:00.003006", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/miimon", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203848.93400: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/miimon', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203848.93402: _low_level_execute_command(): starting 13731 1727203848.93404: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203848.6459277-15163-142166745232190/ > /dev/null 2>&1 && sleep 0' 13731 1727203848.93964: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203848.93967: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203848.93969: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.93972: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203848.93974: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.94029: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.94052: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.94111: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203848.95923: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203848.95989: stderr chunk (state=3): >>><<< 13731 1727203848.95998: stdout chunk (state=3): >>><<< 13731 1727203848.96018: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203848.96026: handler run complete 13731 1727203848.96186: Evaluated conditional (False): False 13731 1727203848.96201: variable 'bond_opt' from source: unknown 13731 1727203848.96210: variable 'result' from source: unknown 13731 1727203848.96225: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203848.96237: attempt loop complete, returning result 13731 1727203848.96257: variable 'bond_opt' from source: unknown 13731 1727203848.96330: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'miimon', 'value': '110'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "miimon", "value": "110" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/miimon" ], "delta": "0:00:00.003006", "end": "2024-09-24 14:50:48.916006", "rc": 0, "start": "2024-09-24 14:50:48.913000" } STDOUT: 110 13731 1727203848.96547: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.96581: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.96584: variable 'omit' from source: magic vars 13731 1727203848.96742: variable 'ansible_distribution_major_version' from source: facts 13731 1727203848.96752: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203848.96781: variable 'omit' from source: magic vars 13731 1727203848.96784: variable 'omit' from source: magic vars 13731 1727203848.96963: variable 'controller_device' from source: play vars 13731 1727203848.96982: variable 'bond_opt' from source: unknown 13731 1727203848.97058: variable 'omit' from source: magic vars 13731 1727203848.97062: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203848.97064: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203848.97066: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203848.97091: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203848.97100: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.97107: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.97222: Set connection var ansible_pipelining to False 13731 1727203848.97277: Set connection var ansible_shell_type to sh 13731 1727203848.97280: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203848.97283: Set connection var ansible_connection to ssh 13731 1727203848.97285: Set connection var ansible_shell_executable to /bin/sh 13731 1727203848.97287: Set connection var ansible_timeout to 10 13731 1727203848.97302: variable 'ansible_shell_executable' from source: unknown 13731 1727203848.97394: variable 'ansible_connection' from source: unknown 13731 1727203848.97397: variable 'ansible_module_compression' from source: unknown 13731 1727203848.97399: variable 'ansible_shell_type' from source: unknown 13731 1727203848.97401: variable 'ansible_shell_executable' from source: unknown 13731 1727203848.97404: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203848.97406: variable 'ansible_pipelining' from source: unknown 13731 1727203848.97408: variable 'ansible_timeout' from source: unknown 13731 1727203848.97410: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203848.97550: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203848.97763: variable 'omit' from source: magic vars 13731 1727203848.97770: starting attempt loop 13731 1727203848.97772: running the handler 13731 1727203848.97865: _low_level_execute_command(): starting 13731 1727203848.97868: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203848.98406: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203848.98414: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203848.98423: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203848.98437: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203848.98447: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203848.98454: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203848.98466: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.98480: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203848.98488: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203848.98494: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203848.98501: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203848.98510: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203848.98523: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203848.98526: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203848.98532: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203848.98541: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203848.98612: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203848.98666: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203848.98704: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.00245: stdout chunk (state=3): >>>/root <<< 13731 1727203849.00382: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.00408: stdout chunk (state=3): >>><<< 13731 1727203849.00411: stderr chunk (state=3): >>><<< 13731 1727203849.00425: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.00437: _low_level_execute_command(): starting 13731 1727203849.00506: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119 `" && echo ansible-tmp-1727203849.0042977-15163-2037497830119="` echo /root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119 `" ) && sleep 0' 13731 1727203849.01043: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.01058: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.01089: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203849.01141: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.01206: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.01224: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.01261: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.01321: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.03143: stdout chunk (state=3): >>>ansible-tmp-1727203849.0042977-15163-2037497830119=/root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119 <<< 13731 1727203849.03312: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.03316: stdout chunk (state=3): >>><<< 13731 1727203849.03319: stderr chunk (state=3): >>><<< 13731 1727203849.03482: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203849.0042977-15163-2037497830119=/root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.03486: variable 'ansible_module_compression' from source: unknown 13731 1727203849.03488: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203849.03490: variable 'ansible_facts' from source: unknown 13731 1727203849.03515: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/AnsiballZ_command.py 13731 1727203849.03741: Sending initial data 13731 1727203849.03750: Sent initial data (154 bytes) 13731 1727203849.04343: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.04359: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.04391: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203849.04499: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.04546: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.04579: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.06387: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203849.06402: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203849.06526: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpfj6l_6r6 /root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/AnsiballZ_command.py <<< 13731 1727203849.06533: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/AnsiballZ_command.py" <<< 13731 1727203849.06565: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpfj6l_6r6" to remote "/root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/AnsiballZ_command.py" <<< 13731 1727203849.07320: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.07324: stdout chunk (state=3): >>><<< 13731 1727203849.07329: stderr chunk (state=3): >>><<< 13731 1727203849.07391: done transferring module to remote 13731 1727203849.07398: _low_level_execute_command(): starting 13731 1727203849.07402: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/ /root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/AnsiballZ_command.py && sleep 0' 13731 1727203849.08094: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.08129: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.08150: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203849.08196: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.08268: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.08337: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.10143: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.10147: stdout chunk (state=3): >>><<< 13731 1727203849.10153: stderr chunk (state=3): >>><<< 13731 1727203849.10254: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.10258: _low_level_execute_command(): starting 13731 1727203849.10260: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/AnsiballZ_command.py && sleep 0' 13731 1727203849.11355: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.11358: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.11361: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203849.11363: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203849.11365: stderr chunk (state=3): >>>debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.11671: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.11681: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.11813: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.26898: stdout chunk (state=3): >>> {"changed": true, "stdout": "64", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/num_grat_arp"], "start": "2024-09-24 14:50:49.265494", "end": "2024-09-24 14:50:49.268490", "delta": "0:00:00.002996", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/num_grat_arp", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203849.28354: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203849.28358: stdout chunk (state=3): >>><<< 13731 1727203849.28491: stderr chunk (state=3): >>><<< 13731 1727203849.28496: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "64", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/num_grat_arp"], "start": "2024-09-24 14:50:49.265494", "end": "2024-09-24 14:50:49.268490", "delta": "0:00:00.002996", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/num_grat_arp", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203849.28498: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/num_grat_arp', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203849.28500: _low_level_execute_command(): starting 13731 1727203849.28503: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203849.0042977-15163-2037497830119/ > /dev/null 2>&1 && sleep 0' 13731 1727203849.29091: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.29100: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.29111: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203849.29125: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203849.29138: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203849.29239: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.29265: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.29585: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.31164: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.31170: stdout chunk (state=3): >>><<< 13731 1727203849.31179: stderr chunk (state=3): >>><<< 13731 1727203849.31195: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.31200: handler run complete 13731 1727203849.31227: Evaluated conditional (False): False 13731 1727203849.31378: variable 'bond_opt' from source: unknown 13731 1727203849.31385: variable 'result' from source: unknown 13731 1727203849.31397: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203849.31408: attempt loop complete, returning result 13731 1727203849.31427: variable 'bond_opt' from source: unknown 13731 1727203849.31497: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'num_grat_arp', 'value': '64'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "num_grat_arp", "value": "64" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/num_grat_arp" ], "delta": "0:00:00.002996", "end": "2024-09-24 14:50:49.268490", "rc": 0, "start": "2024-09-24 14:50:49.265494" } STDOUT: 64 13731 1727203849.31732: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203849.31736: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203849.31738: variable 'omit' from source: magic vars 13731 1727203849.31755: variable 'ansible_distribution_major_version' from source: facts 13731 1727203849.31778: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203849.31782: variable 'omit' from source: magic vars 13731 1727203849.31784: variable 'omit' from source: magic vars 13731 1727203849.31940: variable 'controller_device' from source: play vars 13731 1727203849.31945: variable 'bond_opt' from source: unknown 13731 1727203849.31959: variable 'omit' from source: magic vars 13731 1727203849.31983: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203849.31990: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203849.31998: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203849.32011: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203849.32014: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203849.32017: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203849.32099: Set connection var ansible_pipelining to False 13731 1727203849.32102: Set connection var ansible_shell_type to sh 13731 1727203849.32104: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203849.32157: Set connection var ansible_connection to ssh 13731 1727203849.32160: Set connection var ansible_shell_executable to /bin/sh 13731 1727203849.32164: Set connection var ansible_timeout to 10 13731 1727203849.32169: variable 'ansible_shell_executable' from source: unknown 13731 1727203849.32172: variable 'ansible_connection' from source: unknown 13731 1727203849.32174: variable 'ansible_module_compression' from source: unknown 13731 1727203849.32177: variable 'ansible_shell_type' from source: unknown 13731 1727203849.32179: variable 'ansible_shell_executable' from source: unknown 13731 1727203849.32181: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203849.32183: variable 'ansible_pipelining' from source: unknown 13731 1727203849.32185: variable 'ansible_timeout' from source: unknown 13731 1727203849.32187: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203849.32249: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203849.32265: variable 'omit' from source: magic vars 13731 1727203849.32268: starting attempt loop 13731 1727203849.32270: running the handler 13731 1727203849.32331: _low_level_execute_command(): starting 13731 1727203849.32335: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203849.32849: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.32861: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.32954: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.33005: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.33040: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.34614: stdout chunk (state=3): >>>/root <<< 13731 1727203849.34766: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.34769: stdout chunk (state=3): >>><<< 13731 1727203849.34771: stderr chunk (state=3): >>><<< 13731 1727203849.34788: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.34867: _low_level_execute_command(): starting 13731 1727203849.34870: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192 `" && echo ansible-tmp-1727203849.3479533-15163-61703269238192="` echo /root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192 `" ) && sleep 0' 13731 1727203849.35446: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.35465: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.35482: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203849.35501: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203849.35518: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203849.35545: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.35641: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.35669: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.35733: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.37573: stdout chunk (state=3): >>>ansible-tmp-1727203849.3479533-15163-61703269238192=/root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192 <<< 13731 1727203849.37710: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.37714: stderr chunk (state=3): >>><<< 13731 1727203849.37716: stdout chunk (state=3): >>><<< 13731 1727203849.37880: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203849.3479533-15163-61703269238192=/root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.37884: variable 'ansible_module_compression' from source: unknown 13731 1727203849.37886: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203849.37888: variable 'ansible_facts' from source: unknown 13731 1727203849.37908: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/AnsiballZ_command.py 13731 1727203849.38137: Sending initial data 13731 1727203849.38140: Sent initial data (155 bytes) 13731 1727203849.38681: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.38696: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.38711: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203849.38728: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203849.38743: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203849.38785: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.38812: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203849.38891: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.38917: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.38940: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.38959: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.39028: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.40681: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203849.41153: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203849.41157: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/AnsiballZ_command.py" debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpkmiqojb0" to remote "/root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/AnsiballZ_command.py" <<< 13731 1727203849.41162: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpkmiqojb0 /root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/AnsiballZ_command.py <<< 13731 1727203849.42109: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.42293: stderr chunk (state=3): >>><<< 13731 1727203849.42296: stdout chunk (state=3): >>><<< 13731 1727203849.42331: done transferring module to remote 13731 1727203849.42349: _low_level_execute_command(): starting 13731 1727203849.42372: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/ /root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/AnsiballZ_command.py && sleep 0' 13731 1727203849.43559: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.43646: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.43649: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.43652: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.43958: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.43964: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.45710: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.45714: stdout chunk (state=3): >>><<< 13731 1727203849.45716: stderr chunk (state=3): >>><<< 13731 1727203849.45882: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.45886: _low_level_execute_command(): starting 13731 1727203849.45888: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/AnsiballZ_command.py && sleep 0' 13731 1727203849.47391: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.47418: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.47429: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.47624: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.47856: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.63110: stdout chunk (state=3): >>> {"changed": true, "stdout": "225", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/resend_igmp"], "start": "2024-09-24 14:50:49.626603", "end": "2024-09-24 14:50:49.629795", "delta": "0:00:00.003192", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/resend_igmp", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203849.64490: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.64507: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203849.64567: stderr chunk (state=3): >>><<< 13731 1727203849.64590: stdout chunk (state=3): >>><<< 13731 1727203849.64619: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "225", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/resend_igmp"], "start": "2024-09-24 14:50:49.626603", "end": "2024-09-24 14:50:49.629795", "delta": "0:00:00.003192", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/resend_igmp", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203849.64652: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/resend_igmp', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203849.64682: _low_level_execute_command(): starting 13731 1727203849.64685: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203849.3479533-15163-61703269238192/ > /dev/null 2>&1 && sleep 0' 13731 1727203849.65388: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203849.65393: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.65421: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.65469: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.65493: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.67344: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.67359: stdout chunk (state=3): >>><<< 13731 1727203849.67373: stderr chunk (state=3): >>><<< 13731 1727203849.67400: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.67411: handler run complete 13731 1727203849.67581: Evaluated conditional (False): False 13731 1727203849.67599: variable 'bond_opt' from source: unknown 13731 1727203849.67611: variable 'result' from source: unknown 13731 1727203849.67629: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203849.67644: attempt loop complete, returning result 13731 1727203849.67668: variable 'bond_opt' from source: unknown 13731 1727203849.67738: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'resend_igmp', 'value': '225'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "resend_igmp", "value": "225" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/resend_igmp" ], "delta": "0:00:00.003192", "end": "2024-09-24 14:50:49.629795", "rc": 0, "start": "2024-09-24 14:50:49.626603" } STDOUT: 225 13731 1727203849.68084: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203849.68087: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203849.68090: variable 'omit' from source: magic vars 13731 1727203849.68209: variable 'ansible_distribution_major_version' from source: facts 13731 1727203849.68220: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203849.68229: variable 'omit' from source: magic vars 13731 1727203849.68247: variable 'omit' from source: magic vars 13731 1727203849.68434: variable 'controller_device' from source: play vars 13731 1727203849.68443: variable 'bond_opt' from source: unknown 13731 1727203849.68469: variable 'omit' from source: magic vars 13731 1727203849.68498: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203849.68516: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203849.68532: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203849.68625: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203849.68628: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203849.68633: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203849.68663: Set connection var ansible_pipelining to False 13731 1727203849.68734: Set connection var ansible_shell_type to sh 13731 1727203849.68737: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203849.68739: Set connection var ansible_connection to ssh 13731 1727203849.68743: Set connection var ansible_shell_executable to /bin/sh 13731 1727203849.68749: Set connection var ansible_timeout to 10 13731 1727203849.68751: variable 'ansible_shell_executable' from source: unknown 13731 1727203849.68753: variable 'ansible_connection' from source: unknown 13731 1727203849.68755: variable 'ansible_module_compression' from source: unknown 13731 1727203849.68757: variable 'ansible_shell_type' from source: unknown 13731 1727203849.68759: variable 'ansible_shell_executable' from source: unknown 13731 1727203849.68765: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203849.68777: variable 'ansible_pipelining' from source: unknown 13731 1727203849.68785: variable 'ansible_timeout' from source: unknown 13731 1727203849.68793: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203849.68903: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203849.68951: variable 'omit' from source: magic vars 13731 1727203849.68955: starting attempt loop 13731 1727203849.68957: running the handler 13731 1727203849.68959: _low_level_execute_command(): starting 13731 1727203849.68967: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203849.69607: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.69758: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.69768: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.69771: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.69794: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.71354: stdout chunk (state=3): >>>/root <<< 13731 1727203849.71585: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.71589: stdout chunk (state=3): >>><<< 13731 1727203849.71591: stderr chunk (state=3): >>><<< 13731 1727203849.71594: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.71596: _low_level_execute_command(): starting 13731 1727203849.71598: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523 `" && echo ansible-tmp-1727203849.715095-15163-31066267048523="` echo /root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523 `" ) && sleep 0' 13731 1727203849.72127: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.72136: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.72148: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203849.72164: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203849.72174: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203849.72191: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203849.72235: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.72325: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.72419: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.72445: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.74318: stdout chunk (state=3): >>>ansible-tmp-1727203849.715095-15163-31066267048523=/root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523 <<< 13731 1727203849.74590: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.74593: stdout chunk (state=3): >>><<< 13731 1727203849.74600: stderr chunk (state=3): >>><<< 13731 1727203849.74626: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203849.715095-15163-31066267048523=/root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.74685: variable 'ansible_module_compression' from source: unknown 13731 1727203849.74688: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203849.74691: variable 'ansible_facts' from source: unknown 13731 1727203849.74771: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/AnsiballZ_command.py 13731 1727203849.75198: Sending initial data 13731 1727203849.75201: Sent initial data (154 bytes) 13731 1727203849.75794: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.75807: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.75830: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203849.75856: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203849.75890: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.75906: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203849.75998: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.76010: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.76020: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.76033: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.76099: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.77637: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203849.77681: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203849.77767: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpjq6hrvla /root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/AnsiballZ_command.py <<< 13731 1727203849.77772: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/AnsiballZ_command.py" debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpjq6hrvla" to remote "/root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/AnsiballZ_command.py" <<< 13731 1727203849.78940: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.78944: stdout chunk (state=3): >>><<< 13731 1727203849.78946: stderr chunk (state=3): >>><<< 13731 1727203849.78948: done transferring module to remote 13731 1727203849.78951: _low_level_execute_command(): starting 13731 1727203849.78953: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/ /root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/AnsiballZ_command.py && sleep 0' 13731 1727203849.79563: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203849.79589: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203849.79612: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203849.79632: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203849.79729: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.79849: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.79923: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.79959: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.80068: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203849.81949: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203849.81953: stdout chunk (state=3): >>><<< 13731 1727203849.81955: stderr chunk (state=3): >>><<< 13731 1727203849.81958: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203849.81963: _low_level_execute_command(): starting 13731 1727203849.81965: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/AnsiballZ_command.py && sleep 0' 13731 1727203849.83293: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203849.83377: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203849.83394: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203849.83415: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203849.83485: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203850.99094: stdout chunk (state=3): >>> {"changed": true, "stdout": "0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/updelay"], "start": "2024-09-24 14:50:49.986097", "end": "2024-09-24 14:50:50.990278", "delta": "0:00:01.004181", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/updelay", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203851.00793: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203851.00823: stdout chunk (state=3): >>><<< 13731 1727203851.00826: stderr chunk (state=3): >>><<< 13731 1727203851.00943: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/updelay"], "start": "2024-09-24 14:50:49.986097", "end": "2024-09-24 14:50:50.990278", "delta": "0:00:01.004181", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/updelay", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203851.00951: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/updelay', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203851.00953: _low_level_execute_command(): starting 13731 1727203851.00955: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203849.715095-15163-31066267048523/ > /dev/null 2>&1 && sleep 0' 13731 1727203851.01497: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.01518: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203851.01532: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203851.01547: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203851.01632: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.01665: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.01682: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.01707: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.01774: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.03640: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.03661: stdout chunk (state=3): >>><<< 13731 1727203851.03681: stderr chunk (state=3): >>><<< 13731 1727203851.03881: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.03885: handler run complete 13731 1727203851.03887: Evaluated conditional (False): False 13731 1727203851.03890: variable 'bond_opt' from source: unknown 13731 1727203851.03904: variable 'result' from source: unknown 13731 1727203851.03922: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203851.03938: attempt loop complete, returning result 13731 1727203851.03968: variable 'bond_opt' from source: unknown 13731 1727203851.04050: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'updelay', 'value': '0'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "updelay", "value": "0" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/updelay" ], "delta": "0:00:01.004181", "end": "2024-09-24 14:50:50.990278", "rc": 0, "start": "2024-09-24 14:50:49.986097" } STDOUT: 0 13731 1727203851.04290: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203851.04304: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203851.04318: variable 'omit' from source: magic vars 13731 1727203851.04499: variable 'ansible_distribution_major_version' from source: facts 13731 1727203851.04554: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203851.04558: variable 'omit' from source: magic vars 13731 1727203851.04560: variable 'omit' from source: magic vars 13731 1727203851.04718: variable 'controller_device' from source: play vars 13731 1727203851.04729: variable 'bond_opt' from source: unknown 13731 1727203851.04752: variable 'omit' from source: magic vars 13731 1727203851.04788: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203851.04803: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203851.04813: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203851.04836: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203851.04882: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203851.04885: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203851.04946: Set connection var ansible_pipelining to False 13731 1727203851.04957: Set connection var ansible_shell_type to sh 13731 1727203851.04967: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203851.04982: Set connection var ansible_connection to ssh 13731 1727203851.04998: Set connection var ansible_shell_executable to /bin/sh 13731 1727203851.05044: Set connection var ansible_timeout to 10 13731 1727203851.05047: variable 'ansible_shell_executable' from source: unknown 13731 1727203851.05049: variable 'ansible_connection' from source: unknown 13731 1727203851.05051: variable 'ansible_module_compression' from source: unknown 13731 1727203851.05054: variable 'ansible_shell_type' from source: unknown 13731 1727203851.05056: variable 'ansible_shell_executable' from source: unknown 13731 1727203851.05061: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203851.05071: variable 'ansible_pipelining' from source: unknown 13731 1727203851.05081: variable 'ansible_timeout' from source: unknown 13731 1727203851.05100: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203851.05199: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203851.05210: variable 'omit' from source: magic vars 13731 1727203851.05260: starting attempt loop 13731 1727203851.05263: running the handler 13731 1727203851.05265: _low_level_execute_command(): starting 13731 1727203851.05267: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203851.05861: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.05878: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203851.05918: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203851.05930: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203851.05970: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.06044: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.06073: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.06112: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.06166: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.07729: stdout chunk (state=3): >>>/root <<< 13731 1727203851.07897: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.07901: stdout chunk (state=3): >>><<< 13731 1727203851.07903: stderr chunk (state=3): >>><<< 13731 1727203851.07936: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.08025: _low_level_execute_command(): starting 13731 1727203851.08029: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766 `" && echo ansible-tmp-1727203851.0794084-15163-117089105973766="` echo /root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766 `" ) && sleep 0' 13731 1727203851.08617: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.08631: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203851.08647: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203851.08668: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203851.08689: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203851.08734: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.08801: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.08820: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.08853: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.08914: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.10890: stdout chunk (state=3): >>>ansible-tmp-1727203851.0794084-15163-117089105973766=/root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766 <<< 13731 1727203851.10908: stdout chunk (state=3): >>><<< 13731 1727203851.10927: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.10930: stderr chunk (state=3): >>><<< 13731 1727203851.11002: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203851.0794084-15163-117089105973766=/root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.11008: variable 'ansible_module_compression' from source: unknown 13731 1727203851.11021: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203851.11044: variable 'ansible_facts' from source: unknown 13731 1727203851.11129: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/AnsiballZ_command.py 13731 1727203851.11303: Sending initial data 13731 1727203851.11307: Sent initial data (156 bytes) 13731 1727203851.12100: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.12207: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.12223: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.13719: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203851.13761: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203851.13797: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp48taxpr3 /root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/AnsiballZ_command.py <<< 13731 1727203851.13799: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/AnsiballZ_command.py" <<< 13731 1727203851.14081: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp48taxpr3" to remote "/root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/AnsiballZ_command.py" <<< 13731 1727203851.14607: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.14648: stderr chunk (state=3): >>><<< 13731 1727203851.14662: stdout chunk (state=3): >>><<< 13731 1727203851.14727: done transferring module to remote 13731 1727203851.14749: _low_level_execute_command(): starting 13731 1727203851.14759: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/ /root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/AnsiballZ_command.py && sleep 0' 13731 1727203851.15641: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.15690: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.15764: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.15786: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.15808: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.15877: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.17610: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.17622: stdout chunk (state=3): >>><<< 13731 1727203851.17640: stderr chunk (state=3): >>><<< 13731 1727203851.17728: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.17731: _low_level_execute_command(): starting 13731 1727203851.17733: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/AnsiballZ_command.py && sleep 0' 13731 1727203851.18251: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.18265: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203851.18280: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203851.18294: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203851.18307: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203851.18316: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203851.18391: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.18416: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.18437: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.18458: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.18529: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.33816: stdout chunk (state=3): >>> {"changed": true, "stdout": "1", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/use_carrier"], "start": "2024-09-24 14:50:51.333631", "end": "2024-09-24 14:50:51.336634", "delta": "0:00:00.003003", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/use_carrier", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203851.35281: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203851.35291: stdout chunk (state=3): >>><<< 13731 1727203851.35293: stderr chunk (state=3): >>><<< 13731 1727203851.35425: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "1", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/use_carrier"], "start": "2024-09-24 14:50:51.333631", "end": "2024-09-24 14:50:51.336634", "delta": "0:00:00.003003", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/use_carrier", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203851.35429: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/use_carrier', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203851.35431: _low_level_execute_command(): starting 13731 1727203851.35434: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203851.0794084-15163-117089105973766/ > /dev/null 2>&1 && sleep 0' 13731 1727203851.36021: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.36038: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203851.36055: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203851.36144: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.36186: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.36206: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.36270: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.38132: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.38136: stdout chunk (state=3): >>><<< 13731 1727203851.38139: stderr chunk (state=3): >>><<< 13731 1727203851.38162: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.38169: handler run complete 13731 1727203851.38193: Evaluated conditional (False): False 13731 1727203851.38523: variable 'bond_opt' from source: unknown 13731 1727203851.38526: variable 'result' from source: unknown 13731 1727203851.38529: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203851.38531: attempt loop complete, returning result 13731 1727203851.38533: variable 'bond_opt' from source: unknown 13731 1727203851.38535: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'use_carrier', 'value': '1'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "use_carrier", "value": "1" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/use_carrier" ], "delta": "0:00:00.003003", "end": "2024-09-24 14:50:51.336634", "rc": 0, "start": "2024-09-24 14:50:51.333631" } STDOUT: 1 13731 1727203851.38645: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203851.38649: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203851.38652: variable 'omit' from source: magic vars 13731 1727203851.38834: variable 'ansible_distribution_major_version' from source: facts 13731 1727203851.38838: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203851.38840: variable 'omit' from source: magic vars 13731 1727203851.38842: variable 'omit' from source: magic vars 13731 1727203851.39057: variable 'controller_device' from source: play vars 13731 1727203851.39061: variable 'bond_opt' from source: unknown 13731 1727203851.39063: variable 'omit' from source: magic vars 13731 1727203851.39065: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203851.39076: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203851.39081: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203851.39098: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203851.39101: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203851.39103: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203851.39199: Set connection var ansible_pipelining to False 13731 1727203851.39202: Set connection var ansible_shell_type to sh 13731 1727203851.39205: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203851.39207: Set connection var ansible_connection to ssh 13731 1727203851.39209: Set connection var ansible_shell_executable to /bin/sh 13731 1727203851.39211: Set connection var ansible_timeout to 10 13731 1727203851.39308: variable 'ansible_shell_executable' from source: unknown 13731 1727203851.39311: variable 'ansible_connection' from source: unknown 13731 1727203851.39314: variable 'ansible_module_compression' from source: unknown 13731 1727203851.39316: variable 'ansible_shell_type' from source: unknown 13731 1727203851.39318: variable 'ansible_shell_executable' from source: unknown 13731 1727203851.39320: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203851.39322: variable 'ansible_pipelining' from source: unknown 13731 1727203851.39324: variable 'ansible_timeout' from source: unknown 13731 1727203851.39326: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203851.39352: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203851.39360: variable 'omit' from source: magic vars 13731 1727203851.39368: starting attempt loop 13731 1727203851.39370: running the handler 13731 1727203851.39378: _low_level_execute_command(): starting 13731 1727203851.39380: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203851.40060: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.40232: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.40236: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.40238: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.40270: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.41838: stdout chunk (state=3): >>>/root <<< 13731 1727203851.41958: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.42025: stderr chunk (state=3): >>><<< 13731 1727203851.42127: stdout chunk (state=3): >>><<< 13731 1727203851.42131: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.42134: _low_level_execute_command(): starting 13731 1727203851.42136: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081 `" && echo ansible-tmp-1727203851.420525-15163-148550282328081="` echo /root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081 `" ) && sleep 0' 13731 1727203851.42650: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.42666: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203851.42683: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203851.42782: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.42803: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.42864: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.44707: stdout chunk (state=3): >>>ansible-tmp-1727203851.420525-15163-148550282328081=/root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081 <<< 13731 1727203851.44854: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.44862: stdout chunk (state=3): >>><<< 13731 1727203851.44872: stderr chunk (state=3): >>><<< 13731 1727203851.44898: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203851.420525-15163-148550282328081=/root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.44929: variable 'ansible_module_compression' from source: unknown 13731 1727203851.44966: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203851.44996: variable 'ansible_facts' from source: unknown 13731 1727203851.45077: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/AnsiballZ_command.py 13731 1727203851.45194: Sending initial data 13731 1727203851.45294: Sent initial data (155 bytes) 13731 1727203851.45890: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.45961: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.45980: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.46005: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.46070: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.47582: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203851.47638: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203851.47680: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp10y5_iit /root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/AnsiballZ_command.py <<< 13731 1727203851.47684: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/AnsiballZ_command.py" <<< 13731 1727203851.47733: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp10y5_iit" to remote "/root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/AnsiballZ_command.py" <<< 13731 1727203851.48482: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.48550: stderr chunk (state=3): >>><<< 13731 1727203851.48554: stdout chunk (state=3): >>><<< 13731 1727203851.48563: done transferring module to remote 13731 1727203851.48577: _low_level_execute_command(): starting 13731 1727203851.48587: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/ /root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/AnsiballZ_command.py && sleep 0' 13731 1727203851.49212: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.49236: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203851.49249: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203851.49264: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203851.49342: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203851.49345: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.49385: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203851.49400: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.49418: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.49487: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.51235: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.51239: stdout chunk (state=3): >>><<< 13731 1727203851.51243: stderr chunk (state=3): >>><<< 13731 1727203851.51266: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.51270: _low_level_execute_command(): starting 13731 1727203851.51272: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/AnsiballZ_command.py && sleep 0' 13731 1727203851.52059: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203851.52074: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203851.52081: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.52084: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.52154: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.67683: stdout chunk (state=3): >>> {"changed": true, "stdout": "encap2+3 3", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/xmit_hash_policy"], "start": "2024-09-24 14:50:51.672520", "end": "2024-09-24 14:50:51.675716", "delta": "0:00:00.003196", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/xmit_hash_policy", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203851.69385: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203851.69389: stdout chunk (state=3): >>><<< 13731 1727203851.69392: stderr chunk (state=3): >>><<< 13731 1727203851.69394: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "encap2+3 3", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/xmit_hash_policy"], "start": "2024-09-24 14:50:51.672520", "end": "2024-09-24 14:50:51.675716", "delta": "0:00:00.003196", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/xmit_hash_policy", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203851.69549: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/xmit_hash_policy', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203851.69552: _low_level_execute_command(): starting 13731 1727203851.69555: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203851.420525-15163-148550282328081/ > /dev/null 2>&1 && sleep 0' 13731 1727203851.70147: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203851.70164: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203851.70188: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203851.70222: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203851.70237: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203851.70322: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203851.70355: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203851.70439: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203851.72245: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203851.72249: stdout chunk (state=3): >>><<< 13731 1727203851.72482: stderr chunk (state=3): >>><<< 13731 1727203851.72485: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203851.72488: handler run complete 13731 1727203851.72490: Evaluated conditional (False): False 13731 1727203851.72698: variable 'bond_opt' from source: unknown 13731 1727203851.72704: variable 'result' from source: unknown 13731 1727203851.72717: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203851.72727: attempt loop complete, returning result 13731 1727203851.72745: variable 'bond_opt' from source: unknown 13731 1727203851.72938: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'xmit_hash_policy', 'value': 'encap2+3'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "xmit_hash_policy", "value": "encap2+3" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/xmit_hash_policy" ], "delta": "0:00:00.003196", "end": "2024-09-24 14:50:51.675716", "rc": 0, "start": "2024-09-24 14:50:51.672520" } STDOUT: encap2+3 3 13731 1727203851.73074: dumping result to json 13731 1727203851.73079: done dumping result, returning 13731 1727203851.73290: done running TaskExecutor() for managed-node3/TASK: ** TEST check bond settings [028d2410-947f-82dc-c122-000000000400] 13731 1727203851.73293: sending task result for task 028d2410-947f-82dc-c122-000000000400 13731 1727203851.74324: no more pending results, returning what we have 13731 1727203851.74328: results queue empty 13731 1727203851.74329: checking for any_errors_fatal 13731 1727203851.74334: done checking for any_errors_fatal 13731 1727203851.74335: checking for max_fail_percentage 13731 1727203851.74336: done checking for max_fail_percentage 13731 1727203851.74337: checking to see if all hosts have failed and the running result is not ok 13731 1727203851.74338: done checking to see if all hosts have failed 13731 1727203851.74338: getting the remaining hosts for this loop 13731 1727203851.74340: done getting the remaining hosts for this loop 13731 1727203851.74343: getting the next task for host managed-node3 13731 1727203851.74348: done getting next task for host managed-node3 13731 1727203851.74351: ^ task is: TASK: Include the task 'assert_IPv4_present.yml' 13731 1727203851.74353: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203851.74357: getting variables 13731 1727203851.74358: in VariableManager get_vars() 13731 1727203851.74490: done sending task result for task 028d2410-947f-82dc-c122-000000000400 13731 1727203851.74493: WORKER PROCESS EXITING 13731 1727203851.74504: Calling all_inventory to load vars for managed-node3 13731 1727203851.74506: Calling groups_inventory to load vars for managed-node3 13731 1727203851.74510: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203851.74520: Calling all_plugins_play to load vars for managed-node3 13731 1727203851.74523: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203851.74526: Calling groups_plugins_play to load vars for managed-node3 13731 1727203851.77432: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203851.80595: done with get_vars() 13731 1727203851.80624: done getting variables TASK [Include the task 'assert_IPv4_present.yml'] ****************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml:11 Tuesday 24 September 2024 14:50:51 -0400 (0:00:06.680) 0:00:30.038 ***** 13731 1727203851.80726: entering _queue_task() for managed-node3/include_tasks 13731 1727203851.81495: worker is 1 (out of 1 available) 13731 1727203851.81506: exiting _queue_task() for managed-node3/include_tasks 13731 1727203851.81519: done queuing things up, now waiting for results queue to drain 13731 1727203851.81520: waiting for pending results... 13731 1727203851.82217: running TaskExecutor() for managed-node3/TASK: Include the task 'assert_IPv4_present.yml' 13731 1727203851.82307: in run() - task 028d2410-947f-82dc-c122-000000000402 13731 1727203851.82331: variable 'ansible_search_path' from source: unknown 13731 1727203851.82338: variable 'ansible_search_path' from source: unknown 13731 1727203851.82378: calling self._execute() 13731 1727203851.82644: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203851.82650: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203851.82669: variable 'omit' from source: magic vars 13731 1727203851.83485: variable 'ansible_distribution_major_version' from source: facts 13731 1727203851.83506: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203851.83681: _execute() done 13731 1727203851.83685: dumping result to json 13731 1727203851.83688: done dumping result, returning 13731 1727203851.83691: done running TaskExecutor() for managed-node3/TASK: Include the task 'assert_IPv4_present.yml' [028d2410-947f-82dc-c122-000000000402] 13731 1727203851.83693: sending task result for task 028d2410-947f-82dc-c122-000000000402 13731 1727203851.83981: done sending task result for task 028d2410-947f-82dc-c122-000000000402 13731 1727203851.83985: WORKER PROCESS EXITING 13731 1727203851.84015: no more pending results, returning what we have 13731 1727203851.84021: in VariableManager get_vars() 13731 1727203851.84063: Calling all_inventory to load vars for managed-node3 13731 1727203851.84067: Calling groups_inventory to load vars for managed-node3 13731 1727203851.84072: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203851.84089: Calling all_plugins_play to load vars for managed-node3 13731 1727203851.84092: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203851.84096: Calling groups_plugins_play to load vars for managed-node3 13731 1727203851.86996: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203851.90141: done with get_vars() 13731 1727203851.90170: variable 'ansible_search_path' from source: unknown 13731 1727203851.90172: variable 'ansible_search_path' from source: unknown 13731 1727203851.90185: variable 'item' from source: include params 13731 1727203851.90498: variable 'item' from source: include params 13731 1727203851.90534: we have included files to process 13731 1727203851.90535: generating all_blocks data 13731 1727203851.90538: done generating all_blocks data 13731 1727203851.90543: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml 13731 1727203851.90544: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml 13731 1727203851.90547: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml 13731 1727203851.90986: done processing included file 13731 1727203851.90988: iterating over new_blocks loaded from include file 13731 1727203851.90989: in VariableManager get_vars() 13731 1727203851.91005: done with get_vars() 13731 1727203851.91006: filtering new block on tags 13731 1727203851.91031: done filtering new block on tags 13731 1727203851.91033: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml for managed-node3 13731 1727203851.91038: extending task lists for all hosts with included blocks 13731 1727203851.91432: done extending task lists 13731 1727203851.91433: done processing included files 13731 1727203851.91434: results queue empty 13731 1727203851.91435: checking for any_errors_fatal 13731 1727203851.91448: done checking for any_errors_fatal 13731 1727203851.91449: checking for max_fail_percentage 13731 1727203851.91450: done checking for max_fail_percentage 13731 1727203851.91451: checking to see if all hosts have failed and the running result is not ok 13731 1727203851.91452: done checking to see if all hosts have failed 13731 1727203851.91452: getting the remaining hosts for this loop 13731 1727203851.91454: done getting the remaining hosts for this loop 13731 1727203851.91456: getting the next task for host managed-node3 13731 1727203851.91460: done getting next task for host managed-node3 13731 1727203851.91462: ^ task is: TASK: ** TEST check IPv4 13731 1727203851.91466: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203851.91468: getting variables 13731 1727203851.91469: in VariableManager get_vars() 13731 1727203851.91683: Calling all_inventory to load vars for managed-node3 13731 1727203851.91685: Calling groups_inventory to load vars for managed-node3 13731 1727203851.91688: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203851.91693: Calling all_plugins_play to load vars for managed-node3 13731 1727203851.91696: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203851.91698: Calling groups_plugins_play to load vars for managed-node3 13731 1727203851.93908: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203851.95623: done with get_vars() 13731 1727203851.95652: done getting variables 13731 1727203851.95707: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [** TEST check IPv4] ****************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml:3 Tuesday 24 September 2024 14:50:51 -0400 (0:00:00.150) 0:00:30.188 ***** 13731 1727203851.95744: entering _queue_task() for managed-node3/command 13731 1727203851.96219: worker is 1 (out of 1 available) 13731 1727203851.96231: exiting _queue_task() for managed-node3/command 13731 1727203851.96244: done queuing things up, now waiting for results queue to drain 13731 1727203851.96246: waiting for pending results... 13731 1727203851.96598: running TaskExecutor() for managed-node3/TASK: ** TEST check IPv4 13731 1727203851.96683: in run() - task 028d2410-947f-82dc-c122-000000000631 13731 1727203851.96882: variable 'ansible_search_path' from source: unknown 13731 1727203851.96887: variable 'ansible_search_path' from source: unknown 13731 1727203851.96890: calling self._execute() 13731 1727203851.96893: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203851.96895: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203851.96897: variable 'omit' from source: magic vars 13731 1727203851.97370: variable 'ansible_distribution_major_version' from source: facts 13731 1727203851.97390: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203851.97402: variable 'omit' from source: magic vars 13731 1727203851.97499: variable 'omit' from source: magic vars 13731 1727203851.97689: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203852.01074: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203852.01146: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203852.01187: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203852.01233: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203852.01264: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203852.01356: variable 'interface' from source: include params 13731 1727203852.01367: variable 'controller_device' from source: play vars 13731 1727203852.01444: variable 'controller_device' from source: play vars 13731 1727203852.01474: variable 'omit' from source: magic vars 13731 1727203852.01529: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203852.01715: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203852.01718: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203852.01721: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203852.01723: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203852.01725: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203852.01727: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203852.01729: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203852.01833: Set connection var ansible_pipelining to False 13731 1727203852.01848: Set connection var ansible_shell_type to sh 13731 1727203852.01961: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203852.01979: Set connection var ansible_connection to ssh 13731 1727203852.01991: Set connection var ansible_shell_executable to /bin/sh 13731 1727203852.02001: Set connection var ansible_timeout to 10 13731 1727203852.02028: variable 'ansible_shell_executable' from source: unknown 13731 1727203852.02069: variable 'ansible_connection' from source: unknown 13731 1727203852.02082: variable 'ansible_module_compression' from source: unknown 13731 1727203852.02090: variable 'ansible_shell_type' from source: unknown 13731 1727203852.02098: variable 'ansible_shell_executable' from source: unknown 13731 1727203852.02105: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203852.02112: variable 'ansible_pipelining' from source: unknown 13731 1727203852.02118: variable 'ansible_timeout' from source: unknown 13731 1727203852.02126: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203852.02325: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203852.02328: variable 'omit' from source: magic vars 13731 1727203852.02331: starting attempt loop 13731 1727203852.02333: running the handler 13731 1727203852.02335: _low_level_execute_command(): starting 13731 1727203852.02338: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203852.03092: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203852.03110: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203852.03185: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203852.03239: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.03257: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.03400: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.03547: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.05197: stdout chunk (state=3): >>>/root <<< 13731 1727203852.05338: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.05397: stdout chunk (state=3): >>><<< 13731 1727203852.05401: stderr chunk (state=3): >>><<< 13731 1727203852.05408: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203852.05433: _low_level_execute_command(): starting 13731 1727203852.05493: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017 `" && echo ansible-tmp-1727203852.054211-15787-219184852889017="` echo /root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017 `" ) && sleep 0' 13731 1727203852.06282: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203852.06598: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.06612: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.06627: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.06689: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.08552: stdout chunk (state=3): >>>ansible-tmp-1727203852.054211-15787-219184852889017=/root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017 <<< 13731 1727203852.08722: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.08733: stdout chunk (state=3): >>><<< 13731 1727203852.08749: stderr chunk (state=3): >>><<< 13731 1727203852.08779: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203852.054211-15787-219184852889017=/root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203852.08816: variable 'ansible_module_compression' from source: unknown 13731 1727203852.08879: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203852.09115: variable 'ansible_facts' from source: unknown 13731 1727203852.09206: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/AnsiballZ_command.py 13731 1727203852.09503: Sending initial data 13731 1727203852.09513: Sent initial data (155 bytes) 13731 1727203852.10028: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203852.10044: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203852.10063: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203852.10166: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.10190: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.10202: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.10259: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.11798: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203852.11856: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203852.11911: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp04l7nrx6 /root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/AnsiballZ_command.py <<< 13731 1727203852.11935: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/AnsiballZ_command.py" <<< 13731 1727203852.11969: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory <<< 13731 1727203852.12302: stderr chunk (state=3): >>>debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp04l7nrx6" to remote "/root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/AnsiballZ_command.py" <<< 13731 1727203852.12827: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.12891: stderr chunk (state=3): >>><<< 13731 1727203852.12900: stdout chunk (state=3): >>><<< 13731 1727203852.12956: done transferring module to remote 13731 1727203852.12973: _low_level_execute_command(): starting 13731 1727203852.12986: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/ /root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/AnsiballZ_command.py && sleep 0' 13731 1727203852.13615: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203852.13698: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203852.13739: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.13754: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.13792: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.13852: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.15630: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.15634: stdout chunk (state=3): >>><<< 13731 1727203852.15636: stderr chunk (state=3): >>><<< 13731 1727203852.15667: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203852.15711: _low_level_execute_command(): starting 13731 1727203852.15714: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/AnsiballZ_command.py && sleep 0' 13731 1727203852.16298: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203852.16311: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203852.16328: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203852.16342: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203852.16436: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.16456: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.16483: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.16672: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.32031: stdout chunk (state=3): >>> {"changed": true, "stdout": "18: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet 192.0.2.35/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond\n valid_lft 228sec preferred_lft 228sec", "stderr": "", "rc": 0, "cmd": ["ip", "-4", "a", "s", "nm-bond"], "start": "2024-09-24 14:50:52.316160", "end": "2024-09-24 14:50:52.319680", "delta": "0:00:00.003520", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -4 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203852.33839: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203852.33848: stdout chunk (state=3): >>><<< 13731 1727203852.33851: stderr chunk (state=3): >>><<< 13731 1727203852.33854: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "18: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet 192.0.2.35/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond\n valid_lft 228sec preferred_lft 228sec", "stderr": "", "rc": 0, "cmd": ["ip", "-4", "a", "s", "nm-bond"], "start": "2024-09-24 14:50:52.316160", "end": "2024-09-24 14:50:52.319680", "delta": "0:00:00.003520", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -4 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203852.33888: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip -4 a s nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203852.33895: _low_level_execute_command(): starting 13731 1727203852.33946: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203852.054211-15787-219184852889017/ > /dev/null 2>&1 && sleep 0' 13731 1727203852.35187: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.35215: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.35230: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.35307: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.37129: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.37146: stdout chunk (state=3): >>><<< 13731 1727203852.37158: stderr chunk (state=3): >>><<< 13731 1727203852.37184: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203852.37196: handler run complete 13731 1727203852.37225: Evaluated conditional (False): False 13731 1727203852.37417: variable 'address' from source: include params 13731 1727203852.37428: variable 'result' from source: set_fact 13731 1727203852.37448: Evaluated conditional (address in result.stdout): True 13731 1727203852.37470: attempt loop complete, returning result 13731 1727203852.37482: _execute() done 13731 1727203852.37584: dumping result to json 13731 1727203852.37587: done dumping result, returning 13731 1727203852.37589: done running TaskExecutor() for managed-node3/TASK: ** TEST check IPv4 [028d2410-947f-82dc-c122-000000000631] 13731 1727203852.37592: sending task result for task 028d2410-947f-82dc-c122-000000000631 13731 1727203852.37669: done sending task result for task 028d2410-947f-82dc-c122-000000000631 13731 1727203852.37673: WORKER PROCESS EXITING ok: [managed-node3] => { "attempts": 1, "changed": false, "cmd": [ "ip", "-4", "a", "s", "nm-bond" ], "delta": "0:00:00.003520", "end": "2024-09-24 14:50:52.319680", "rc": 0, "start": "2024-09-24 14:50:52.316160" } STDOUT: 18: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000 inet 192.0.2.35/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond valid_lft 228sec preferred_lft 228sec 13731 1727203852.37779: no more pending results, returning what we have 13731 1727203852.37784: results queue empty 13731 1727203852.37785: checking for any_errors_fatal 13731 1727203852.37787: done checking for any_errors_fatal 13731 1727203852.37788: checking for max_fail_percentage 13731 1727203852.37790: done checking for max_fail_percentage 13731 1727203852.37791: checking to see if all hosts have failed and the running result is not ok 13731 1727203852.37791: done checking to see if all hosts have failed 13731 1727203852.37792: getting the remaining hosts for this loop 13731 1727203852.37795: done getting the remaining hosts for this loop 13731 1727203852.37798: getting the next task for host managed-node3 13731 1727203852.37808: done getting next task for host managed-node3 13731 1727203852.37811: ^ task is: TASK: Include the task 'assert_IPv6_present.yml' 13731 1727203852.37815: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203852.37821: getting variables 13731 1727203852.37822: in VariableManager get_vars() 13731 1727203852.37857: Calling all_inventory to load vars for managed-node3 13731 1727203852.37863: Calling groups_inventory to load vars for managed-node3 13731 1727203852.37867: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203852.37881: Calling all_plugins_play to load vars for managed-node3 13731 1727203852.37884: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203852.37888: Calling groups_plugins_play to load vars for managed-node3 13731 1727203852.39717: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203852.41246: done with get_vars() 13731 1727203852.41271: done getting variables TASK [Include the task 'assert_IPv6_present.yml'] ****************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml:16 Tuesday 24 September 2024 14:50:52 -0400 (0:00:00.456) 0:00:30.644 ***** 13731 1727203852.41357: entering _queue_task() for managed-node3/include_tasks 13731 1727203852.41658: worker is 1 (out of 1 available) 13731 1727203852.41671: exiting _queue_task() for managed-node3/include_tasks 13731 1727203852.41884: done queuing things up, now waiting for results queue to drain 13731 1727203852.41886: waiting for pending results... 13731 1727203852.42013: running TaskExecutor() for managed-node3/TASK: Include the task 'assert_IPv6_present.yml' 13731 1727203852.42068: in run() - task 028d2410-947f-82dc-c122-000000000403 13731 1727203852.42092: variable 'ansible_search_path' from source: unknown 13731 1727203852.42099: variable 'ansible_search_path' from source: unknown 13731 1727203852.42145: calling self._execute() 13731 1727203852.42255: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203852.42267: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203852.42284: variable 'omit' from source: magic vars 13731 1727203852.42763: variable 'ansible_distribution_major_version' from source: facts 13731 1727203852.42767: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203852.42769: _execute() done 13731 1727203852.42772: dumping result to json 13731 1727203852.42774: done dumping result, returning 13731 1727203852.42779: done running TaskExecutor() for managed-node3/TASK: Include the task 'assert_IPv6_present.yml' [028d2410-947f-82dc-c122-000000000403] 13731 1727203852.42781: sending task result for task 028d2410-947f-82dc-c122-000000000403 13731 1727203852.42851: done sending task result for task 028d2410-947f-82dc-c122-000000000403 13731 1727203852.42854: WORKER PROCESS EXITING 13731 1727203852.42893: no more pending results, returning what we have 13731 1727203852.42900: in VariableManager get_vars() 13731 1727203852.42938: Calling all_inventory to load vars for managed-node3 13731 1727203852.42941: Calling groups_inventory to load vars for managed-node3 13731 1727203852.42945: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203852.42960: Calling all_plugins_play to load vars for managed-node3 13731 1727203852.42963: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203852.42967: Calling groups_plugins_play to load vars for managed-node3 13731 1727203852.44436: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203852.45967: done with get_vars() 13731 1727203852.45994: variable 'ansible_search_path' from source: unknown 13731 1727203852.45995: variable 'ansible_search_path' from source: unknown 13731 1727203852.46005: variable 'item' from source: include params 13731 1727203852.46112: variable 'item' from source: include params 13731 1727203852.46146: we have included files to process 13731 1727203852.46147: generating all_blocks data 13731 1727203852.46149: done generating all_blocks data 13731 1727203852.46154: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml 13731 1727203852.46156: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml 13731 1727203852.46158: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml 13731 1727203852.46399: done processing included file 13731 1727203852.46401: iterating over new_blocks loaded from include file 13731 1727203852.46402: in VariableManager get_vars() 13731 1727203852.46420: done with get_vars() 13731 1727203852.46421: filtering new block on tags 13731 1727203852.46448: done filtering new block on tags 13731 1727203852.46451: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml for managed-node3 13731 1727203852.46457: extending task lists for all hosts with included blocks 13731 1727203852.46778: done extending task lists 13731 1727203852.46780: done processing included files 13731 1727203852.46780: results queue empty 13731 1727203852.46781: checking for any_errors_fatal 13731 1727203852.46786: done checking for any_errors_fatal 13731 1727203852.46786: checking for max_fail_percentage 13731 1727203852.46787: done checking for max_fail_percentage 13731 1727203852.46788: checking to see if all hosts have failed and the running result is not ok 13731 1727203852.46789: done checking to see if all hosts have failed 13731 1727203852.46790: getting the remaining hosts for this loop 13731 1727203852.46791: done getting the remaining hosts for this loop 13731 1727203852.46793: getting the next task for host managed-node3 13731 1727203852.46796: done getting next task for host managed-node3 13731 1727203852.46798: ^ task is: TASK: ** TEST check IPv6 13731 1727203852.46802: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203852.46803: getting variables 13731 1727203852.46804: in VariableManager get_vars() 13731 1727203852.46812: Calling all_inventory to load vars for managed-node3 13731 1727203852.46814: Calling groups_inventory to load vars for managed-node3 13731 1727203852.46816: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203852.46821: Calling all_plugins_play to load vars for managed-node3 13731 1727203852.46823: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203852.46825: Calling groups_plugins_play to load vars for managed-node3 13731 1727203852.47960: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203852.49353: done with get_vars() 13731 1727203852.49374: done getting variables 13731 1727203852.49420: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [** TEST check IPv6] ****************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml:3 Tuesday 24 September 2024 14:50:52 -0400 (0:00:00.080) 0:00:30.725 ***** 13731 1727203852.49451: entering _queue_task() for managed-node3/command 13731 1727203852.49789: worker is 1 (out of 1 available) 13731 1727203852.49801: exiting _queue_task() for managed-node3/command 13731 1727203852.49814: done queuing things up, now waiting for results queue to drain 13731 1727203852.49816: waiting for pending results... 13731 1727203852.50195: running TaskExecutor() for managed-node3/TASK: ** TEST check IPv6 13731 1727203852.50228: in run() - task 028d2410-947f-82dc-c122-000000000652 13731 1727203852.50246: variable 'ansible_search_path' from source: unknown 13731 1727203852.50252: variable 'ansible_search_path' from source: unknown 13731 1727203852.50294: calling self._execute() 13731 1727203852.50388: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203852.50403: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203852.50415: variable 'omit' from source: magic vars 13731 1727203852.50772: variable 'ansible_distribution_major_version' from source: facts 13731 1727203852.50790: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203852.50799: variable 'omit' from source: magic vars 13731 1727203852.50854: variable 'omit' from source: magic vars 13731 1727203852.51011: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203852.53060: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203852.53137: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203852.53181: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203852.53224: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203852.53257: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203852.53349: variable 'controller_device' from source: play vars 13731 1727203852.53380: variable 'omit' from source: magic vars 13731 1727203852.53417: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203852.53456: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203852.53484: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203852.53544: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203852.53548: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203852.53561: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203852.53570: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203852.53582: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203852.53691: Set connection var ansible_pipelining to False 13731 1727203852.53703: Set connection var ansible_shell_type to sh 13731 1727203852.53761: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203852.53764: Set connection var ansible_connection to ssh 13731 1727203852.53766: Set connection var ansible_shell_executable to /bin/sh 13731 1727203852.53769: Set connection var ansible_timeout to 10 13731 1727203852.53771: variable 'ansible_shell_executable' from source: unknown 13731 1727203852.53773: variable 'ansible_connection' from source: unknown 13731 1727203852.53780: variable 'ansible_module_compression' from source: unknown 13731 1727203852.53788: variable 'ansible_shell_type' from source: unknown 13731 1727203852.53795: variable 'ansible_shell_executable' from source: unknown 13731 1727203852.53801: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203852.53810: variable 'ansible_pipelining' from source: unknown 13731 1727203852.53818: variable 'ansible_timeout' from source: unknown 13731 1727203852.53828: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203852.53982: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203852.53985: variable 'omit' from source: magic vars 13731 1727203852.53988: starting attempt loop 13731 1727203852.53990: running the handler 13731 1727203852.53998: _low_level_execute_command(): starting 13731 1727203852.54009: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203852.54688: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203852.54736: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203852.54749: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203852.54813: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.54854: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.54922: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.56578: stdout chunk (state=3): >>>/root <<< 13731 1727203852.56736: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.56740: stdout chunk (state=3): >>><<< 13731 1727203852.56742: stderr chunk (state=3): >>><<< 13731 1727203852.56782: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203852.56789: _low_level_execute_command(): starting 13731 1727203852.56875: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312 `" && echo ansible-tmp-1727203852.5676749-15876-99501734275312="` echo /root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312 `" ) && sleep 0' 13731 1727203852.57494: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203852.57513: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.57529: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.57552: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.57625: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.59685: stdout chunk (state=3): >>>ansible-tmp-1727203852.5676749-15876-99501734275312=/root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312 <<< 13731 1727203852.59690: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.59693: stdout chunk (state=3): >>><<< 13731 1727203852.59696: stderr chunk (state=3): >>><<< 13731 1727203852.59715: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203852.5676749-15876-99501734275312=/root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203852.59749: variable 'ansible_module_compression' from source: unknown 13731 1727203852.59949: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203852.59992: variable 'ansible_facts' from source: unknown 13731 1727203852.60095: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/AnsiballZ_command.py 13731 1727203852.60452: Sending initial data 13731 1727203852.60456: Sent initial data (155 bytes) 13731 1727203852.60982: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203852.61094: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203852.61106: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.61118: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.61131: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.61195: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.62697: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203852.62839: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203852.62842: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp9amtqnf8 /root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/AnsiballZ_command.py <<< 13731 1727203852.62845: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/AnsiballZ_command.py" <<< 13731 1727203852.62848: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp9amtqnf8" to remote "/root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/AnsiballZ_command.py" <<< 13731 1727203852.64250: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.64255: stdout chunk (state=3): >>><<< 13731 1727203852.64257: stderr chunk (state=3): >>><<< 13731 1727203852.64469: done transferring module to remote 13731 1727203852.64473: _low_level_execute_command(): starting 13731 1727203852.64481: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/ /root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/AnsiballZ_command.py && sleep 0' 13731 1727203852.65628: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203852.65632: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203852.65694: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203852.65785: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203852.65934: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.65948: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.67724: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.68527: stderr chunk (state=3): >>><<< 13731 1727203852.68531: stdout chunk (state=3): >>><<< 13731 1727203852.68534: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203852.68537: _low_level_execute_command(): starting 13731 1727203852.68539: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/AnsiballZ_command.py && sleep 0' 13731 1727203852.69715: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203852.69739: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203852.69750: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203852.69766: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203852.69795: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203852.69802: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203852.70188: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.70401: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.70433: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.85811: stdout chunk (state=3): >>> {"changed": true, "stdout": "18: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet6 2001:db8::b9/128 scope global dynamic noprefixroute \n valid_lft 228sec preferred_lft 228sec\n inet6 2001:db8::7c30:ebff:fe61:ecfc/64 scope global dynamic noprefixroute \n valid_lft 1792sec preferred_lft 1792sec\n inet6 fe80::7c30:ebff:fe61:ecfc/64 scope link noprefixroute \n valid_lft forever preferred_lft forever", "stderr": "", "rc": 0, "cmd": ["ip", "-6", "a", "s", "nm-bond"], "start": "2024-09-24 14:50:52.852884", "end": "2024-09-24 14:50:52.856581", "delta": "0:00:00.003697", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -6 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203852.87346: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203852.87350: stdout chunk (state=3): >>><<< 13731 1727203852.87353: stderr chunk (state=3): >>><<< 13731 1727203852.87380: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "18: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet6 2001:db8::b9/128 scope global dynamic noprefixroute \n valid_lft 228sec preferred_lft 228sec\n inet6 2001:db8::7c30:ebff:fe61:ecfc/64 scope global dynamic noprefixroute \n valid_lft 1792sec preferred_lft 1792sec\n inet6 fe80::7c30:ebff:fe61:ecfc/64 scope link noprefixroute \n valid_lft forever preferred_lft forever", "stderr": "", "rc": 0, "cmd": ["ip", "-6", "a", "s", "nm-bond"], "start": "2024-09-24 14:50:52.852884", "end": "2024-09-24 14:50:52.856581", "delta": "0:00:00.003697", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -6 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203852.87430: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip -6 a s nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203852.87493: _low_level_execute_command(): starting 13731 1727203852.87504: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203852.5676749-15876-99501734275312/ > /dev/null 2>&1 && sleep 0' 13731 1727203852.88778: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203852.88797: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203852.88812: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203852.88849: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203852.89065: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203852.89085: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203852.89112: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203852.89297: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203852.91106: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203852.91247: stderr chunk (state=3): >>><<< 13731 1727203852.91250: stdout chunk (state=3): >>><<< 13731 1727203852.91270: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203852.91551: handler run complete 13731 1727203852.91554: Evaluated conditional (False): False 13731 1727203852.91785: variable 'address' from source: include params 13731 1727203852.91796: variable 'result' from source: set_fact 13731 1727203852.91817: Evaluated conditional (address in result.stdout): True 13731 1727203852.91834: attempt loop complete, returning result 13731 1727203852.91842: _execute() done 13731 1727203852.91848: dumping result to json 13731 1727203852.91887: done dumping result, returning 13731 1727203852.91900: done running TaskExecutor() for managed-node3/TASK: ** TEST check IPv6 [028d2410-947f-82dc-c122-000000000652] 13731 1727203852.91911: sending task result for task 028d2410-947f-82dc-c122-000000000652 13731 1727203852.92292: done sending task result for task 028d2410-947f-82dc-c122-000000000652 13731 1727203852.92295: WORKER PROCESS EXITING ok: [managed-node3] => { "attempts": 1, "changed": false, "cmd": [ "ip", "-6", "a", "s", "nm-bond" ], "delta": "0:00:00.003697", "end": "2024-09-24 14:50:52.856581", "rc": 0, "start": "2024-09-24 14:50:52.852884" } STDOUT: 18: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000 inet6 2001:db8::b9/128 scope global dynamic noprefixroute valid_lft 228sec preferred_lft 228sec inet6 2001:db8::7c30:ebff:fe61:ecfc/64 scope global dynamic noprefixroute valid_lft 1792sec preferred_lft 1792sec inet6 fe80::7c30:ebff:fe61:ecfc/64 scope link noprefixroute valid_lft forever preferred_lft forever 13731 1727203852.92380: no more pending results, returning what we have 13731 1727203852.92384: results queue empty 13731 1727203852.92385: checking for any_errors_fatal 13731 1727203852.92387: done checking for any_errors_fatal 13731 1727203852.92388: checking for max_fail_percentage 13731 1727203852.92390: done checking for max_fail_percentage 13731 1727203852.92391: checking to see if all hosts have failed and the running result is not ok 13731 1727203852.92392: done checking to see if all hosts have failed 13731 1727203852.92393: getting the remaining hosts for this loop 13731 1727203852.92395: done getting the remaining hosts for this loop 13731 1727203852.92399: getting the next task for host managed-node3 13731 1727203852.92410: done getting next task for host managed-node3 13731 1727203852.92414: ^ task is: TASK: Conditional asserts 13731 1727203852.92416: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203852.92421: getting variables 13731 1727203852.92423: in VariableManager get_vars() 13731 1727203852.92456: Calling all_inventory to load vars for managed-node3 13731 1727203852.92459: Calling groups_inventory to load vars for managed-node3 13731 1727203852.92466: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203852.92680: Calling all_plugins_play to load vars for managed-node3 13731 1727203852.92684: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203852.92689: Calling groups_plugins_play to load vars for managed-node3 13731 1727203852.96222: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203853.00031: done with get_vars() 13731 1727203853.00059: done getting variables TASK [Conditional asserts] ***************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:42 Tuesday 24 September 2024 14:50:53 -0400 (0:00:00.508) 0:00:31.234 ***** 13731 1727203853.00306: entering _queue_task() for managed-node3/include_tasks 13731 1727203853.01056: worker is 1 (out of 1 available) 13731 1727203853.01070: exiting _queue_task() for managed-node3/include_tasks 13731 1727203853.01086: done queuing things up, now waiting for results queue to drain 13731 1727203853.01088: waiting for pending results... 13731 1727203853.01695: running TaskExecutor() for managed-node3/TASK: Conditional asserts 13731 1727203853.01895: in run() - task 028d2410-947f-82dc-c122-00000000008e 13731 1727203853.01916: variable 'ansible_search_path' from source: unknown 13731 1727203853.02038: variable 'ansible_search_path' from source: unknown 13731 1727203853.02683: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203853.07187: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203853.07390: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203853.07433: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203853.07591: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203853.07680: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203853.07881: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203853.07885: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203853.07887: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203853.08044: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203853.08069: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203853.08551: dumping result to json 13731 1727203853.08555: done dumping result, returning 13731 1727203853.08557: done running TaskExecutor() for managed-node3/TASK: Conditional asserts [028d2410-947f-82dc-c122-00000000008e] 13731 1727203853.08563: sending task result for task 028d2410-947f-82dc-c122-00000000008e 13731 1727203853.08643: done sending task result for task 028d2410-947f-82dc-c122-00000000008e 13731 1727203853.08646: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "skipped_reason": "No items in the list" } 13731 1727203853.08711: no more pending results, returning what we have 13731 1727203853.08715: results queue empty 13731 1727203853.08716: checking for any_errors_fatal 13731 1727203853.08728: done checking for any_errors_fatal 13731 1727203853.08729: checking for max_fail_percentage 13731 1727203853.08731: done checking for max_fail_percentage 13731 1727203853.08732: checking to see if all hosts have failed and the running result is not ok 13731 1727203853.08733: done checking to see if all hosts have failed 13731 1727203853.08734: getting the remaining hosts for this loop 13731 1727203853.08736: done getting the remaining hosts for this loop 13731 1727203853.08740: getting the next task for host managed-node3 13731 1727203853.08747: done getting next task for host managed-node3 13731 1727203853.08751: ^ task is: TASK: Success in test '{{ lsr_description }}' 13731 1727203853.08754: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203853.08759: getting variables 13731 1727203853.08763: in VariableManager get_vars() 13731 1727203853.08803: Calling all_inventory to load vars for managed-node3 13731 1727203853.08807: Calling groups_inventory to load vars for managed-node3 13731 1727203853.08811: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203853.08823: Calling all_plugins_play to load vars for managed-node3 13731 1727203853.08826: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203853.08829: Calling groups_plugins_play to load vars for managed-node3 13731 1727203853.11984: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203853.14954: done with get_vars() 13731 1727203853.14991: done getting variables 13731 1727203853.15055: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203853.15386: variable 'lsr_description' from source: include params TASK [Success in test 'Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.'] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:47 Tuesday 24 September 2024 14:50:53 -0400 (0:00:00.151) 0:00:31.385 ***** 13731 1727203853.15418: entering _queue_task() for managed-node3/debug 13731 1727203853.16393: worker is 1 (out of 1 available) 13731 1727203853.16407: exiting _queue_task() for managed-node3/debug 13731 1727203853.16419: done queuing things up, now waiting for results queue to drain 13731 1727203853.16420: waiting for pending results... 13731 1727203853.17298: running TaskExecutor() for managed-node3/TASK: Success in test 'Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.' 13731 1727203853.17308: in run() - task 028d2410-947f-82dc-c122-00000000008f 13731 1727203853.17311: variable 'ansible_search_path' from source: unknown 13731 1727203853.17314: variable 'ansible_search_path' from source: unknown 13731 1727203853.17884: calling self._execute() 13731 1727203853.17890: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203853.17894: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203853.17897: variable 'omit' from source: magic vars 13731 1727203853.18704: variable 'ansible_distribution_major_version' from source: facts 13731 1727203853.18725: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203853.18738: variable 'omit' from source: magic vars 13731 1727203853.18817: variable 'omit' from source: magic vars 13731 1727203853.19031: variable 'lsr_description' from source: include params 13731 1727203853.19058: variable 'omit' from source: magic vars 13731 1727203853.19184: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203853.19225: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203853.19303: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203853.19327: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203853.19368: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203853.19497: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203853.19506: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203853.19515: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203853.19742: Set connection var ansible_pipelining to False 13731 1727203853.19754: Set connection var ansible_shell_type to sh 13731 1727203853.19765: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203853.19774: Set connection var ansible_connection to ssh 13731 1727203853.19791: Set connection var ansible_shell_executable to /bin/sh 13731 1727203853.19802: Set connection var ansible_timeout to 10 13731 1727203853.19980: variable 'ansible_shell_executable' from source: unknown 13731 1727203853.19983: variable 'ansible_connection' from source: unknown 13731 1727203853.19985: variable 'ansible_module_compression' from source: unknown 13731 1727203853.19988: variable 'ansible_shell_type' from source: unknown 13731 1727203853.19990: variable 'ansible_shell_executable' from source: unknown 13731 1727203853.19992: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203853.19993: variable 'ansible_pipelining' from source: unknown 13731 1727203853.19996: variable 'ansible_timeout' from source: unknown 13731 1727203853.19998: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203853.20283: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203853.20287: variable 'omit' from source: magic vars 13731 1727203853.20289: starting attempt loop 13731 1727203853.20291: running the handler 13731 1727203853.20294: handler run complete 13731 1727203853.20411: attempt loop complete, returning result 13731 1727203853.20608: _execute() done 13731 1727203853.20612: dumping result to json 13731 1727203853.20615: done dumping result, returning 13731 1727203853.20617: done running TaskExecutor() for managed-node3/TASK: Success in test 'Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.' [028d2410-947f-82dc-c122-00000000008f] 13731 1727203853.20621: sending task result for task 028d2410-947f-82dc-c122-00000000008f ok: [managed-node3] => {} MSG: +++++ Success in test 'Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.' +++++ 13731 1727203853.20741: no more pending results, returning what we have 13731 1727203853.20746: results queue empty 13731 1727203853.20747: checking for any_errors_fatal 13731 1727203853.20753: done checking for any_errors_fatal 13731 1727203853.20754: checking for max_fail_percentage 13731 1727203853.20756: done checking for max_fail_percentage 13731 1727203853.20757: checking to see if all hosts have failed and the running result is not ok 13731 1727203853.20758: done checking to see if all hosts have failed 13731 1727203853.20759: getting the remaining hosts for this loop 13731 1727203853.20763: done getting the remaining hosts for this loop 13731 1727203853.20768: getting the next task for host managed-node3 13731 1727203853.20982: done getting next task for host managed-node3 13731 1727203853.20987: ^ task is: TASK: Cleanup 13731 1727203853.20990: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203853.20996: getting variables 13731 1727203853.20998: in VariableManager get_vars() 13731 1727203853.21030: Calling all_inventory to load vars for managed-node3 13731 1727203853.21033: Calling groups_inventory to load vars for managed-node3 13731 1727203853.21036: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203853.21048: Calling all_plugins_play to load vars for managed-node3 13731 1727203853.21051: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203853.21056: Calling groups_plugins_play to load vars for managed-node3 13731 1727203853.21884: done sending task result for task 028d2410-947f-82dc-c122-00000000008f 13731 1727203853.21888: WORKER PROCESS EXITING 13731 1727203853.36267: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203853.39791: done with get_vars() 13731 1727203853.39822: done getting variables TASK [Cleanup] ***************************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:66 Tuesday 24 September 2024 14:50:53 -0400 (0:00:00.244) 0:00:31.630 ***** 13731 1727203853.39905: entering _queue_task() for managed-node3/include_tasks 13731 1727203853.40644: worker is 1 (out of 1 available) 13731 1727203853.40658: exiting _queue_task() for managed-node3/include_tasks 13731 1727203853.40671: done queuing things up, now waiting for results queue to drain 13731 1727203853.40673: waiting for pending results... 13731 1727203853.41172: running TaskExecutor() for managed-node3/TASK: Cleanup 13731 1727203853.41593: in run() - task 028d2410-947f-82dc-c122-000000000093 13731 1727203853.41598: variable 'ansible_search_path' from source: unknown 13731 1727203853.41602: variable 'ansible_search_path' from source: unknown 13731 1727203853.41604: variable 'lsr_cleanup' from source: include params 13731 1727203853.41981: variable 'lsr_cleanup' from source: include params 13731 1727203853.42206: variable 'omit' from source: magic vars 13731 1727203853.42506: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203853.42521: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203853.42536: variable 'omit' from source: magic vars 13731 1727203853.43182: variable 'ansible_distribution_major_version' from source: facts 13731 1727203853.43188: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203853.43192: variable 'item' from source: unknown 13731 1727203853.43194: variable 'item' from source: unknown 13731 1727203853.43382: variable 'item' from source: unknown 13731 1727203853.43423: variable 'item' from source: unknown 13731 1727203853.43783: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203853.43787: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203853.43790: variable 'omit' from source: magic vars 13731 1727203853.44067: variable 'ansible_distribution_major_version' from source: facts 13731 1727203853.44117: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203853.44217: variable 'item' from source: unknown 13731 1727203853.44281: variable 'item' from source: unknown 13731 1727203853.44382: variable 'item' from source: unknown 13731 1727203853.44653: variable 'item' from source: unknown 13731 1727203853.44717: dumping result to json 13731 1727203853.44720: done dumping result, returning 13731 1727203853.44724: done running TaskExecutor() for managed-node3/TASK: Cleanup [028d2410-947f-82dc-c122-000000000093] 13731 1727203853.44726: sending task result for task 028d2410-947f-82dc-c122-000000000093 13731 1727203853.44894: done sending task result for task 028d2410-947f-82dc-c122-000000000093 13731 1727203853.44898: WORKER PROCESS EXITING 13731 1727203853.44924: no more pending results, returning what we have 13731 1727203853.44930: in VariableManager get_vars() 13731 1727203853.44970: Calling all_inventory to load vars for managed-node3 13731 1727203853.44973: Calling groups_inventory to load vars for managed-node3 13731 1727203853.44978: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203853.44995: Calling all_plugins_play to load vars for managed-node3 13731 1727203853.44998: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203853.45001: Calling groups_plugins_play to load vars for managed-node3 13731 1727203853.47766: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203853.50868: done with get_vars() 13731 1727203853.50894: variable 'ansible_search_path' from source: unknown 13731 1727203853.50895: variable 'ansible_search_path' from source: unknown 13731 1727203853.50938: variable 'ansible_search_path' from source: unknown 13731 1727203853.50939: variable 'ansible_search_path' from source: unknown 13731 1727203853.50966: we have included files to process 13731 1727203853.50968: generating all_blocks data 13731 1727203853.50970: done generating all_blocks data 13731 1727203853.50974: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml 13731 1727203853.51180: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml 13731 1727203853.51185: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml 13731 1727203853.51641: in VariableManager get_vars() 13731 1727203853.51663: done with get_vars() 13731 1727203853.51670: variable 'omit' from source: magic vars 13731 1727203853.51712: variable 'omit' from source: magic vars 13731 1727203853.51762: in VariableManager get_vars() 13731 1727203853.51774: done with get_vars() 13731 1727203853.51803: in VariableManager get_vars() 13731 1727203853.51821: done with get_vars() 13731 1727203853.51861: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/defaults/main.yml 13731 1727203853.52236: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/meta/main.yml 13731 1727203853.52522: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml 13731 1727203853.53120: in VariableManager get_vars() 13731 1727203853.53139: done with get_vars() redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13731 1727203853.57119: done processing included file 13731 1727203853.57121: iterating over new_blocks loaded from include file 13731 1727203853.57122: in VariableManager get_vars() 13731 1727203853.57423: done with get_vars() 13731 1727203853.57425: filtering new block on tags 13731 1727203853.58157: done filtering new block on tags 13731 1727203853.58161: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml for managed-node3 => (item=tasks/cleanup_bond_profile+device.yml) 13731 1727203853.58166: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml 13731 1727203853.58167: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml 13731 1727203853.58170: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml 13731 1727203853.58961: done processing included file 13731 1727203853.58963: iterating over new_blocks loaded from include file 13731 1727203853.58964: in VariableManager get_vars() 13731 1727203853.58982: done with get_vars() 13731 1727203853.58985: filtering new block on tags 13731 1727203853.59015: done filtering new block on tags 13731 1727203853.59017: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml for managed-node3 => (item=tasks/remove_test_interfaces_with_dhcp.yml) 13731 1727203853.59021: extending task lists for all hosts with included blocks 13731 1727203853.62754: done extending task lists 13731 1727203853.62756: done processing included files 13731 1727203853.62756: results queue empty 13731 1727203853.62757: checking for any_errors_fatal 13731 1727203853.62762: done checking for any_errors_fatal 13731 1727203853.62763: checking for max_fail_percentage 13731 1727203853.62764: done checking for max_fail_percentage 13731 1727203853.62765: checking to see if all hosts have failed and the running result is not ok 13731 1727203853.62766: done checking to see if all hosts have failed 13731 1727203853.62766: getting the remaining hosts for this loop 13731 1727203853.62768: done getting the remaining hosts for this loop 13731 1727203853.62770: getting the next task for host managed-node3 13731 1727203853.62980: done getting next task for host managed-node3 13731 1727203853.62990: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13731 1727203853.62994: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203853.63005: getting variables 13731 1727203853.63007: in VariableManager get_vars() 13731 1727203853.63025: Calling all_inventory to load vars for managed-node3 13731 1727203853.63028: Calling groups_inventory to load vars for managed-node3 13731 1727203853.63030: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203853.63036: Calling all_plugins_play to load vars for managed-node3 13731 1727203853.63038: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203853.63041: Calling groups_plugins_play to load vars for managed-node3 13731 1727203853.65736: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203853.68888: done with get_vars() 13731 1727203853.68921: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:4 Tuesday 24 September 2024 14:50:53 -0400 (0:00:00.290) 0:00:31.921 ***** 13731 1727203853.69006: entering _queue_task() for managed-node3/include_tasks 13731 1727203853.69769: worker is 1 (out of 1 available) 13731 1727203853.69783: exiting _queue_task() for managed-node3/include_tasks 13731 1727203853.69796: done queuing things up, now waiting for results queue to drain 13731 1727203853.69797: waiting for pending results... 13731 1727203853.70795: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13731 1727203853.71436: in run() - task 028d2410-947f-82dc-c122-000000000693 13731 1727203853.71441: variable 'ansible_search_path' from source: unknown 13731 1727203853.71444: variable 'ansible_search_path' from source: unknown 13731 1727203853.71446: calling self._execute() 13731 1727203853.71970: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203853.71978: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203853.71982: variable 'omit' from source: magic vars 13731 1727203853.73782: variable 'ansible_distribution_major_version' from source: facts 13731 1727203853.73786: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203853.73789: _execute() done 13731 1727203853.73792: dumping result to json 13731 1727203853.73794: done dumping result, returning 13731 1727203853.73796: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role [028d2410-947f-82dc-c122-000000000693] 13731 1727203853.73799: sending task result for task 028d2410-947f-82dc-c122-000000000693 13731 1727203853.73874: done sending task result for task 028d2410-947f-82dc-c122-000000000693 13731 1727203853.73922: no more pending results, returning what we have 13731 1727203853.73929: in VariableManager get_vars() 13731 1727203853.73980: Calling all_inventory to load vars for managed-node3 13731 1727203853.73983: Calling groups_inventory to load vars for managed-node3 13731 1727203853.73986: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203853.74000: Calling all_plugins_play to load vars for managed-node3 13731 1727203853.74003: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203853.74006: Calling groups_plugins_play to load vars for managed-node3 13731 1727203853.74684: WORKER PROCESS EXITING 13731 1727203853.76935: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203853.80062: done with get_vars() 13731 1727203853.80088: variable 'ansible_search_path' from source: unknown 13731 1727203853.80090: variable 'ansible_search_path' from source: unknown 13731 1727203853.80133: we have included files to process 13731 1727203853.80134: generating all_blocks data 13731 1727203853.80136: done generating all_blocks data 13731 1727203853.80137: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203853.80138: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203853.80141: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203853.81199: done processing included file 13731 1727203853.81201: iterating over new_blocks loaded from include file 13731 1727203853.81203: in VariableManager get_vars() 13731 1727203853.81229: done with get_vars() 13731 1727203853.81231: filtering new block on tags 13731 1727203853.81264: done filtering new block on tags 13731 1727203853.81267: in VariableManager get_vars() 13731 1727203853.81594: done with get_vars() 13731 1727203853.81596: filtering new block on tags 13731 1727203853.81643: done filtering new block on tags 13731 1727203853.81646: in VariableManager get_vars() 13731 1727203853.81669: done with get_vars() 13731 1727203853.81671: filtering new block on tags 13731 1727203853.81719: done filtering new block on tags 13731 1727203853.81721: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml for managed-node3 13731 1727203853.81727: extending task lists for all hosts with included blocks 13731 1727203853.85129: done extending task lists 13731 1727203853.85131: done processing included files 13731 1727203853.85132: results queue empty 13731 1727203853.85132: checking for any_errors_fatal 13731 1727203853.85136: done checking for any_errors_fatal 13731 1727203853.85137: checking for max_fail_percentage 13731 1727203853.85138: done checking for max_fail_percentage 13731 1727203853.85139: checking to see if all hosts have failed and the running result is not ok 13731 1727203853.85140: done checking to see if all hosts have failed 13731 1727203853.85141: getting the remaining hosts for this loop 13731 1727203853.85142: done getting the remaining hosts for this loop 13731 1727203853.85144: getting the next task for host managed-node3 13731 1727203853.85150: done getting next task for host managed-node3 13731 1727203853.85153: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13731 1727203853.85157: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203853.85167: getting variables 13731 1727203853.85168: in VariableManager get_vars() 13731 1727203853.85187: Calling all_inventory to load vars for managed-node3 13731 1727203853.85189: Calling groups_inventory to load vars for managed-node3 13731 1727203853.85191: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203853.85197: Calling all_plugins_play to load vars for managed-node3 13731 1727203853.85199: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203853.85202: Calling groups_plugins_play to load vars for managed-node3 13731 1727203853.87663: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203853.90878: done with get_vars() 13731 1727203853.90901: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role are present] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:3 Tuesday 24 September 2024 14:50:53 -0400 (0:00:00.221) 0:00:32.142 ***** 13731 1727203853.91188: entering _queue_task() for managed-node3/setup 13731 1727203853.91740: worker is 1 (out of 1 available) 13731 1727203853.91753: exiting _queue_task() for managed-node3/setup 13731 1727203853.91765: done queuing things up, now waiting for results queue to drain 13731 1727203853.91767: waiting for pending results... 13731 1727203853.92265: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13731 1727203853.92642: in run() - task 028d2410-947f-82dc-c122-0000000007c9 13731 1727203853.92665: variable 'ansible_search_path' from source: unknown 13731 1727203853.92943: variable 'ansible_search_path' from source: unknown 13731 1727203853.92948: calling self._execute() 13731 1727203853.92982: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203853.92996: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203853.93011: variable 'omit' from source: magic vars 13731 1727203853.93752: variable 'ansible_distribution_major_version' from source: facts 13731 1727203853.93813: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203853.94285: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203853.98459: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203853.98664: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203853.98736: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203853.99032: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203853.99035: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203853.99250: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203853.99254: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203853.99257: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203853.99260: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203853.99371: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203853.99432: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203853.99464: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203853.99690: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203853.99693: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203853.99695: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203853.99965: variable '__network_required_facts' from source: role '' defaults 13731 1727203854.00233: variable 'ansible_facts' from source: unknown 13731 1727203854.01593: Evaluated conditional (__network_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 13731 1727203854.01603: when evaluation is False, skipping this task 13731 1727203854.01612: _execute() done 13731 1727203854.01729: dumping result to json 13731 1727203854.01732: done dumping result, returning 13731 1727203854.01735: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present [028d2410-947f-82dc-c122-0000000007c9] 13731 1727203854.01737: sending task result for task 028d2410-947f-82dc-c122-0000000007c9 13731 1727203854.01804: done sending task result for task 028d2410-947f-82dc-c122-0000000007c9 13731 1727203854.01807: WORKER PROCESS EXITING skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203854.01878: no more pending results, returning what we have 13731 1727203854.01883: results queue empty 13731 1727203854.01884: checking for any_errors_fatal 13731 1727203854.01886: done checking for any_errors_fatal 13731 1727203854.01886: checking for max_fail_percentage 13731 1727203854.01888: done checking for max_fail_percentage 13731 1727203854.01889: checking to see if all hosts have failed and the running result is not ok 13731 1727203854.01890: done checking to see if all hosts have failed 13731 1727203854.01890: getting the remaining hosts for this loop 13731 1727203854.01892: done getting the remaining hosts for this loop 13731 1727203854.01895: getting the next task for host managed-node3 13731 1727203854.01906: done getting next task for host managed-node3 13731 1727203854.01910: ^ task is: TASK: fedora.linux_system_roles.network : Check if system is ostree 13731 1727203854.01915: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203854.01931: getting variables 13731 1727203854.01933: in VariableManager get_vars() 13731 1727203854.01972: Calling all_inventory to load vars for managed-node3 13731 1727203854.01975: Calling groups_inventory to load vars for managed-node3 13731 1727203854.01980: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203854.01991: Calling all_plugins_play to load vars for managed-node3 13731 1727203854.01994: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203854.02004: Calling groups_plugins_play to load vars for managed-node3 13731 1727203854.05005: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203854.07970: done with get_vars() 13731 1727203854.08207: done getting variables TASK [fedora.linux_system_roles.network : Check if system is ostree] *********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:12 Tuesday 24 September 2024 14:50:54 -0400 (0:00:00.171) 0:00:32.314 ***** 13731 1727203854.08312: entering _queue_task() for managed-node3/stat 13731 1727203854.09068: worker is 1 (out of 1 available) 13731 1727203854.09085: exiting _queue_task() for managed-node3/stat 13731 1727203854.09099: done queuing things up, now waiting for results queue to drain 13731 1727203854.09101: waiting for pending results... 13731 1727203854.09696: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if system is ostree 13731 1727203854.10138: in run() - task 028d2410-947f-82dc-c122-0000000007cb 13731 1727203854.10143: variable 'ansible_search_path' from source: unknown 13731 1727203854.10146: variable 'ansible_search_path' from source: unknown 13731 1727203854.10150: calling self._execute() 13731 1727203854.10305: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203854.10318: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203854.10334: variable 'omit' from source: magic vars 13731 1727203854.11181: variable 'ansible_distribution_major_version' from source: facts 13731 1727203854.11185: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203854.11552: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203854.12000: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203854.12133: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203854.12166: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203854.12234: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203854.12385: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203854.12608: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203854.12635: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203854.12667: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203854.12764: variable '__network_is_ostree' from source: set_fact 13731 1727203854.12835: Evaluated conditional (not __network_is_ostree is defined): False 13731 1727203854.12844: when evaluation is False, skipping this task 13731 1727203854.12853: _execute() done 13731 1727203854.12860: dumping result to json 13731 1727203854.12867: done dumping result, returning 13731 1727203854.12880: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if system is ostree [028d2410-947f-82dc-c122-0000000007cb] 13731 1727203854.12891: sending task result for task 028d2410-947f-82dc-c122-0000000007cb skipping: [managed-node3] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13731 1727203854.13040: no more pending results, returning what we have 13731 1727203854.13044: results queue empty 13731 1727203854.13046: checking for any_errors_fatal 13731 1727203854.13056: done checking for any_errors_fatal 13731 1727203854.13057: checking for max_fail_percentage 13731 1727203854.13059: done checking for max_fail_percentage 13731 1727203854.13060: checking to see if all hosts have failed and the running result is not ok 13731 1727203854.13061: done checking to see if all hosts have failed 13731 1727203854.13062: getting the remaining hosts for this loop 13731 1727203854.13064: done getting the remaining hosts for this loop 13731 1727203854.13068: getting the next task for host managed-node3 13731 1727203854.13079: done getting next task for host managed-node3 13731 1727203854.13083: ^ task is: TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13731 1727203854.13090: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203854.13109: getting variables 13731 1727203854.13111: in VariableManager get_vars() 13731 1727203854.13153: Calling all_inventory to load vars for managed-node3 13731 1727203854.13156: Calling groups_inventory to load vars for managed-node3 13731 1727203854.13159: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203854.13170: Calling all_plugins_play to load vars for managed-node3 13731 1727203854.13173: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203854.13480: Calling groups_plugins_play to load vars for managed-node3 13731 1727203854.14382: done sending task result for task 028d2410-947f-82dc-c122-0000000007cb 13731 1727203854.14386: WORKER PROCESS EXITING 13731 1727203854.16507: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203854.19544: done with get_vars() 13731 1727203854.19580: done getting variables 13731 1727203854.19642: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Set flag to indicate system is ostree] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:17 Tuesday 24 September 2024 14:50:54 -0400 (0:00:00.115) 0:00:32.429 ***** 13731 1727203854.19890: entering _queue_task() for managed-node3/set_fact 13731 1727203854.20652: worker is 1 (out of 1 available) 13731 1727203854.20666: exiting _queue_task() for managed-node3/set_fact 13731 1727203854.20681: done queuing things up, now waiting for results queue to drain 13731 1727203854.20683: waiting for pending results... 13731 1727203854.21388: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13731 1727203854.22055: in run() - task 028d2410-947f-82dc-c122-0000000007cc 13731 1727203854.22074: variable 'ansible_search_path' from source: unknown 13731 1727203854.22079: variable 'ansible_search_path' from source: unknown 13731 1727203854.22121: calling self._execute() 13731 1727203854.22556: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203854.22561: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203854.22577: variable 'omit' from source: magic vars 13731 1727203854.23893: variable 'ansible_distribution_major_version' from source: facts 13731 1727203854.23914: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203854.24244: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203854.24825: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203854.24933: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203854.25033: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203854.25251: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203854.25327: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203854.25363: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203854.25397: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203854.25456: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203854.25683: variable '__network_is_ostree' from source: set_fact 13731 1727203854.25791: Evaluated conditional (not __network_is_ostree is defined): False 13731 1727203854.25794: when evaluation is False, skipping this task 13731 1727203854.25797: _execute() done 13731 1727203854.25799: dumping result to json 13731 1727203854.25802: done dumping result, returning 13731 1727203854.25805: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree [028d2410-947f-82dc-c122-0000000007cc] 13731 1727203854.25808: sending task result for task 028d2410-947f-82dc-c122-0000000007cc skipping: [managed-node3] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13731 1727203854.26075: no more pending results, returning what we have 13731 1727203854.26082: results queue empty 13731 1727203854.26083: checking for any_errors_fatal 13731 1727203854.26092: done checking for any_errors_fatal 13731 1727203854.26093: checking for max_fail_percentage 13731 1727203854.26095: done checking for max_fail_percentage 13731 1727203854.26096: checking to see if all hosts have failed and the running result is not ok 13731 1727203854.26096: done checking to see if all hosts have failed 13731 1727203854.26097: getting the remaining hosts for this loop 13731 1727203854.26100: done getting the remaining hosts for this loop 13731 1727203854.26103: getting the next task for host managed-node3 13731 1727203854.26116: done getting next task for host managed-node3 13731 1727203854.26120: ^ task is: TASK: fedora.linux_system_roles.network : Check which services are running 13731 1727203854.26127: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203854.26148: getting variables 13731 1727203854.26151: in VariableManager get_vars() 13731 1727203854.26195: Calling all_inventory to load vars for managed-node3 13731 1727203854.26198: Calling groups_inventory to load vars for managed-node3 13731 1727203854.26200: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203854.26212: Calling all_plugins_play to load vars for managed-node3 13731 1727203854.26215: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203854.26218: Calling groups_plugins_play to load vars for managed-node3 13731 1727203854.26812: done sending task result for task 028d2410-947f-82dc-c122-0000000007cc 13731 1727203854.26816: WORKER PROCESS EXITING 13731 1727203854.29443: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203854.32787: done with get_vars() 13731 1727203854.32820: done getting variables TASK [fedora.linux_system_roles.network : Check which services are running] **** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 Tuesday 24 September 2024 14:50:54 -0400 (0:00:00.131) 0:00:32.561 ***** 13731 1727203854.33043: entering _queue_task() for managed-node3/service_facts 13731 1727203854.33881: worker is 1 (out of 1 available) 13731 1727203854.33895: exiting _queue_task() for managed-node3/service_facts 13731 1727203854.33908: done queuing things up, now waiting for results queue to drain 13731 1727203854.33909: waiting for pending results... 13731 1727203854.34487: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which services are running 13731 1727203854.34782: in run() - task 028d2410-947f-82dc-c122-0000000007ce 13731 1727203854.34786: variable 'ansible_search_path' from source: unknown 13731 1727203854.34789: variable 'ansible_search_path' from source: unknown 13731 1727203854.34822: calling self._execute() 13731 1727203854.35154: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203854.35158: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203854.35162: variable 'omit' from source: magic vars 13731 1727203854.35981: variable 'ansible_distribution_major_version' from source: facts 13731 1727203854.35985: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203854.35987: variable 'omit' from source: magic vars 13731 1727203854.36057: variable 'omit' from source: magic vars 13731 1727203854.36163: variable 'omit' from source: magic vars 13731 1727203854.36281: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203854.36321: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203854.36570: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203854.36574: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203854.36579: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203854.36582: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203854.36584: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203854.36587: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203854.36764: Set connection var ansible_pipelining to False 13731 1727203854.36794: Set connection var ansible_shell_type to sh 13731 1727203854.36906: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203854.36980: Set connection var ansible_connection to ssh 13731 1727203854.36984: Set connection var ansible_shell_executable to /bin/sh 13731 1727203854.36986: Set connection var ansible_timeout to 10 13731 1727203854.36989: variable 'ansible_shell_executable' from source: unknown 13731 1727203854.36991: variable 'ansible_connection' from source: unknown 13731 1727203854.36993: variable 'ansible_module_compression' from source: unknown 13731 1727203854.36995: variable 'ansible_shell_type' from source: unknown 13731 1727203854.36997: variable 'ansible_shell_executable' from source: unknown 13731 1727203854.37000: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203854.37002: variable 'ansible_pipelining' from source: unknown 13731 1727203854.37004: variable 'ansible_timeout' from source: unknown 13731 1727203854.37006: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203854.37383: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203854.37464: variable 'omit' from source: magic vars 13731 1727203854.37474: starting attempt loop 13731 1727203854.37483: running the handler 13731 1727203854.37616: _low_level_execute_command(): starting 13731 1727203854.37619: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203854.39094: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203854.39160: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203854.39313: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203854.39348: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203854.39398: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203854.41078: stdout chunk (state=3): >>>/root <<< 13731 1727203854.41217: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203854.41482: stdout chunk (state=3): >>><<< 13731 1727203854.41485: stderr chunk (state=3): >>><<< 13731 1727203854.41489: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203854.41491: _low_level_execute_command(): starting 13731 1727203854.41494: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204 `" && echo ansible-tmp-1727203854.4136932-15987-46588184362204="` echo /root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204 `" ) && sleep 0' 13731 1727203854.42824: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203854.42924: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203854.43044: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203854.43069: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203854.43142: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203854.45030: stdout chunk (state=3): >>>ansible-tmp-1727203854.4136932-15987-46588184362204=/root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204 <<< 13731 1727203854.45354: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203854.45358: stdout chunk (state=3): >>><<< 13731 1727203854.45364: stderr chunk (state=3): >>><<< 13731 1727203854.45386: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203854.4136932-15987-46588184362204=/root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203854.45438: variable 'ansible_module_compression' from source: unknown 13731 1727203854.45601: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.service_facts-ZIP_DEFLATED 13731 1727203854.45640: variable 'ansible_facts' from source: unknown 13731 1727203854.46115: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/AnsiballZ_service_facts.py 13731 1727203854.46604: Sending initial data 13731 1727203854.46608: Sent initial data (161 bytes) 13731 1727203854.47664: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203854.47668: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203854.47684: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203854.47940: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203854.48169: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203854.49568: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203854.49634: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203854.49725: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp5hvt2fud /root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/AnsiballZ_service_facts.py <<< 13731 1727203854.49729: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/AnsiballZ_service_facts.py" <<< 13731 1727203854.49762: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp5hvt2fud" to remote "/root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/AnsiballZ_service_facts.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/AnsiballZ_service_facts.py" <<< 13731 1727203854.51181: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203854.51270: stderr chunk (state=3): >>><<< 13731 1727203854.51273: stdout chunk (state=3): >>><<< 13731 1727203854.51293: done transferring module to remote 13731 1727203854.51487: _low_level_execute_command(): starting 13731 1727203854.51490: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/ /root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/AnsiballZ_service_facts.py && sleep 0' 13731 1727203854.52684: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203854.52755: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203854.52908: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203854.52931: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203854.53094: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203854.54838: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203854.54865: stdout chunk (state=3): >>><<< 13731 1727203854.54869: stderr chunk (state=3): >>><<< 13731 1727203854.55063: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203854.55067: _low_level_execute_command(): starting 13731 1727203854.55069: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/AnsiballZ_service_facts.py && sleep 0' 13731 1727203854.56152: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203854.56277: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203854.56382: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203854.56438: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203854.56812: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203854.56817: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203856.08266: stdout chunk (state=3): >>> {"ansible_facts": {"services": {"audit-rules.service": {"name": "audit-rules.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dm-event.service": {"name": "dm-event.service", "state": "stopped", "status": "static", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "st<<< 13731 1727203856.08337: stdout chunk (state=3): >>>opped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-lvmpolld.service": {"name": "lvm2-lvmpolld.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-monitor.service": {"name": "lvm2-monitor.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@dm_mod.service": {"name": "modprobe@dm_mod.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@efi_pstore.service": {"name": "modprobe@efi_pstore.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@loop.service": {"name": "modprobe@loop.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "pcscd.service": {"name": "pcscd.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ssh-host-keys-migration.service": {"name": "ssh-host-keys-migration.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-battery-check.service": {"name": "systemd-battery-check.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-confext.service": {"name": "systemd-confext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-clear.service": {"name": "systemd-hibernate-clear.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-resume.service": {"name": "systemd-hibernate-resume.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-oomd.service": {"name": "systemd-oomd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-quotacheck-root.service": {"name": "systemd-quotacheck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-soft-reboot.service": {"name": "systemd-soft-reboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev-early.service": {"name": "systemd-tmpfiles-setup-dev-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup-early.service": {"name": "systemd-tpm2-setup-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup.service": {"name": "systemd-tpm2-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-load-credentials.service": {"name": "systemd-udev-load-credentials.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "stat<<< 13731 1727203856.08367: stdout chunk (state=3): >>>us": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "blk-availability.service": {"name": "blk-availability.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "capsule@.service": {"name": "capsule@.service", "state": "unknown", "status": "static", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd.service": {"name": "dhcpcd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd@.service": {"name": "dhcpcd@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fips-crypto-policy-overlay.service": {"name": "fips-crypto-policy-overlay.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fsidd.service": {"name": "fsidd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "lvm-devices-import.service": {"name": "lvm-devices-import.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "polkit.service": {"name": "polkit.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon-root.service": {"name": "quotaon-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "quotaon@.service": {"name": "quotaon@.service", "state": "unknown", "status": "static", "source": "systemd"}, "rpmdb-migrate.service": {"name": "rpmdb-migrate.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-bootctl@.service": {"name": "systemd-bootctl@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-creds@.service": {"name": "systemd-creds@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald-sync@.service": {"name": "systemd-journald-sync@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrextend@.service": {"name": "systemd-pcrextend@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrlock-file-system.service": {"name": "systemd-pcrlock-file-system.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-code.service": {"name": "systemd-pcrlock-firmware-code.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-config.service": {"name": "systemd-pcrlock-firmware-config.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-machine-id.service": {"name": "systemd-pcrlock-machine-id.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-make-policy.service": {"name": "systemd-pcrlock-make-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-authority.service": {"name": "systemd-pcrlock-secureboot-authority.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-policy.service": {"name": "systemd-pcrlock-secureboot-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock@.service": {"name": "systemd-pcrlock@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-quotacheck@.service": {"name": "systemd-quotacheck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysext@.service": {"name": "systemd-sysext@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} <<< 13731 1727203856.09994: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203856.09998: stdout chunk (state=3): >>><<< 13731 1727203856.10001: stderr chunk (state=3): >>><<< 13731 1727203856.10005: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"services": {"audit-rules.service": {"name": "audit-rules.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dm-event.service": {"name": "dm-event.service", "state": "stopped", "status": "static", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-lvmpolld.service": {"name": "lvm2-lvmpolld.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-monitor.service": {"name": "lvm2-monitor.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@dm_mod.service": {"name": "modprobe@dm_mod.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@efi_pstore.service": {"name": "modprobe@efi_pstore.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@loop.service": {"name": "modprobe@loop.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "pcscd.service": {"name": "pcscd.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ssh-host-keys-migration.service": {"name": "ssh-host-keys-migration.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-battery-check.service": {"name": "systemd-battery-check.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-confext.service": {"name": "systemd-confext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-clear.service": {"name": "systemd-hibernate-clear.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-resume.service": {"name": "systemd-hibernate-resume.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-oomd.service": {"name": "systemd-oomd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-quotacheck-root.service": {"name": "systemd-quotacheck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-soft-reboot.service": {"name": "systemd-soft-reboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev-early.service": {"name": "systemd-tmpfiles-setup-dev-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup-early.service": {"name": "systemd-tpm2-setup-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup.service": {"name": "systemd-tpm2-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-load-credentials.service": {"name": "systemd-udev-load-credentials.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "blk-availability.service": {"name": "blk-availability.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "capsule@.service": {"name": "capsule@.service", "state": "unknown", "status": "static", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd.service": {"name": "dhcpcd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd@.service": {"name": "dhcpcd@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fips-crypto-policy-overlay.service": {"name": "fips-crypto-policy-overlay.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fsidd.service": {"name": "fsidd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "lvm-devices-import.service": {"name": "lvm-devices-import.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "polkit.service": {"name": "polkit.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon-root.service": {"name": "quotaon-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "quotaon@.service": {"name": "quotaon@.service", "state": "unknown", "status": "static", "source": "systemd"}, "rpmdb-migrate.service": {"name": "rpmdb-migrate.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-bootctl@.service": {"name": "systemd-bootctl@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-creds@.service": {"name": "systemd-creds@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald-sync@.service": {"name": "systemd-journald-sync@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrextend@.service": {"name": "systemd-pcrextend@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrlock-file-system.service": {"name": "systemd-pcrlock-file-system.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-code.service": {"name": "systemd-pcrlock-firmware-code.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-config.service": {"name": "systemd-pcrlock-firmware-config.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-machine-id.service": {"name": "systemd-pcrlock-machine-id.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-make-policy.service": {"name": "systemd-pcrlock-make-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-authority.service": {"name": "systemd-pcrlock-secureboot-authority.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-policy.service": {"name": "systemd-pcrlock-secureboot-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock@.service": {"name": "systemd-pcrlock@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-quotacheck@.service": {"name": "systemd-quotacheck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysext@.service": {"name": "systemd-sysext@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203856.11086: done with _execute_module (service_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'service_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203856.11103: _low_level_execute_command(): starting 13731 1727203856.11113: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203854.4136932-15987-46588184362204/ > /dev/null 2>&1 && sleep 0' 13731 1727203856.11811: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203856.11826: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203856.11848: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203856.11874: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203856.11893: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203856.11905: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203856.11970: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203856.12022: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203856.12047: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203856.12091: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203856.12154: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203856.14084: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203856.14118: stderr chunk (state=3): >>><<< 13731 1727203856.14122: stdout chunk (state=3): >>><<< 13731 1727203856.14181: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203856.14206: handler run complete 13731 1727203856.14694: variable 'ansible_facts' from source: unknown 13731 1727203856.14902: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203856.15421: variable 'ansible_facts' from source: unknown 13731 1727203856.15805: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203856.16456: attempt loop complete, returning result 13731 1727203856.16462: _execute() done 13731 1727203856.16465: dumping result to json 13731 1727203856.16538: done dumping result, returning 13731 1727203856.16681: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which services are running [028d2410-947f-82dc-c122-0000000007ce] 13731 1727203856.16684: sending task result for task 028d2410-947f-82dc-c122-0000000007ce 13731 1727203856.18634: done sending task result for task 028d2410-947f-82dc-c122-0000000007ce 13731 1727203856.18637: WORKER PROCESS EXITING ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203856.18738: no more pending results, returning what we have 13731 1727203856.18741: results queue empty 13731 1727203856.18742: checking for any_errors_fatal 13731 1727203856.18744: done checking for any_errors_fatal 13731 1727203856.18745: checking for max_fail_percentage 13731 1727203856.18753: done checking for max_fail_percentage 13731 1727203856.18754: checking to see if all hosts have failed and the running result is not ok 13731 1727203856.18755: done checking to see if all hosts have failed 13731 1727203856.18756: getting the remaining hosts for this loop 13731 1727203856.18757: done getting the remaining hosts for this loop 13731 1727203856.18763: getting the next task for host managed-node3 13731 1727203856.18769: done getting next task for host managed-node3 13731 1727203856.18772: ^ task is: TASK: fedora.linux_system_roles.network : Check which packages are installed 13731 1727203856.18780: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203856.18790: getting variables 13731 1727203856.18791: in VariableManager get_vars() 13731 1727203856.18818: Calling all_inventory to load vars for managed-node3 13731 1727203856.18820: Calling groups_inventory to load vars for managed-node3 13731 1727203856.18822: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203856.18831: Calling all_plugins_play to load vars for managed-node3 13731 1727203856.18837: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203856.18840: Calling groups_plugins_play to load vars for managed-node3 13731 1727203856.21278: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203856.24426: done with get_vars() 13731 1727203856.24464: done getting variables TASK [fedora.linux_system_roles.network : Check which packages are installed] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 Tuesday 24 September 2024 14:50:56 -0400 (0:00:01.915) 0:00:34.476 ***** 13731 1727203856.24579: entering _queue_task() for managed-node3/package_facts 13731 1727203856.25024: worker is 1 (out of 1 available) 13731 1727203856.25038: exiting _queue_task() for managed-node3/package_facts 13731 1727203856.25118: done queuing things up, now waiting for results queue to drain 13731 1727203856.25120: waiting for pending results... 13731 1727203856.25355: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which packages are installed 13731 1727203856.25551: in run() - task 028d2410-947f-82dc-c122-0000000007cf 13731 1727203856.25556: variable 'ansible_search_path' from source: unknown 13731 1727203856.25561: variable 'ansible_search_path' from source: unknown 13731 1727203856.25591: calling self._execute() 13731 1727203856.25688: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203856.25767: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203856.25781: variable 'omit' from source: magic vars 13731 1727203856.26625: variable 'ansible_distribution_major_version' from source: facts 13731 1727203856.26629: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203856.26631: variable 'omit' from source: magic vars 13731 1727203856.26866: variable 'omit' from source: magic vars 13731 1727203856.27283: variable 'omit' from source: magic vars 13731 1727203856.27286: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203856.27289: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203856.27413: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203856.27881: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203856.27884: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203856.27887: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203856.27889: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203856.27892: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203856.27894: Set connection var ansible_pipelining to False 13731 1727203856.27896: Set connection var ansible_shell_type to sh 13731 1727203856.27897: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203856.27899: Set connection var ansible_connection to ssh 13731 1727203856.27901: Set connection var ansible_shell_executable to /bin/sh 13731 1727203856.27903: Set connection var ansible_timeout to 10 13731 1727203856.28066: variable 'ansible_shell_executable' from source: unknown 13731 1727203856.28135: variable 'ansible_connection' from source: unknown 13731 1727203856.28347: variable 'ansible_module_compression' from source: unknown 13731 1727203856.28350: variable 'ansible_shell_type' from source: unknown 13731 1727203856.28352: variable 'ansible_shell_executable' from source: unknown 13731 1727203856.28354: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203856.28356: variable 'ansible_pipelining' from source: unknown 13731 1727203856.28358: variable 'ansible_timeout' from source: unknown 13731 1727203856.28362: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203856.28843: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203856.28905: variable 'omit' from source: magic vars 13731 1727203856.28919: starting attempt loop 13731 1727203856.28930: running the handler 13731 1727203856.29139: _low_level_execute_command(): starting 13731 1727203856.29152: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203856.30959: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203856.31377: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203856.31512: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203856.33066: stdout chunk (state=3): >>>/root <<< 13731 1727203856.33223: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203856.33227: stdout chunk (state=3): >>><<< 13731 1727203856.33230: stderr chunk (state=3): >>><<< 13731 1727203856.33257: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203856.33284: _low_level_execute_command(): starting 13731 1727203856.33452: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935 `" && echo ansible-tmp-1727203856.3326702-16108-248345217445935="` echo /root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935 `" ) && sleep 0' 13731 1727203856.35186: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203856.35574: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203856.35996: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203856.37469: stdout chunk (state=3): >>>ansible-tmp-1727203856.3326702-16108-248345217445935=/root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935 <<< 13731 1727203856.37863: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203856.37867: stdout chunk (state=3): >>><<< 13731 1727203856.37869: stderr chunk (state=3): >>><<< 13731 1727203856.37878: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203856.3326702-16108-248345217445935=/root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203856.37914: variable 'ansible_module_compression' from source: unknown 13731 1727203856.38054: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.package_facts-ZIP_DEFLATED 13731 1727203856.38057: variable 'ansible_facts' from source: unknown 13731 1727203856.38525: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/AnsiballZ_package_facts.py 13731 1727203856.38898: Sending initial data 13731 1727203856.38902: Sent initial data (162 bytes) 13731 1727203856.40184: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203856.40188: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203856.40394: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203856.40398: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203856.40514: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203856.42038: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203856.42064: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203856.42098: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp517dg4i5 /root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/AnsiballZ_package_facts.py <<< 13731 1727203856.42162: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/AnsiballZ_package_facts.py" <<< 13731 1727203856.42179: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp517dg4i5" to remote "/root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/AnsiballZ_package_facts.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/AnsiballZ_package_facts.py" <<< 13731 1727203856.45169: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203856.45255: stderr chunk (state=3): >>><<< 13731 1727203856.45265: stdout chunk (state=3): >>><<< 13731 1727203856.45482: done transferring module to remote 13731 1727203856.45486: _low_level_execute_command(): starting 13731 1727203856.45492: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/ /root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/AnsiballZ_package_facts.py && sleep 0' 13731 1727203856.46668: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203856.46884: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203856.46993: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203856.47422: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203856.49269: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203856.49296: stdout chunk (state=3): >>><<< 13731 1727203856.49401: stderr chunk (state=3): >>><<< 13731 1727203856.49781: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203856.49792: _low_level_execute_command(): starting 13731 1727203856.49794: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/AnsiballZ_package_facts.py && sleep 0' 13731 1727203856.51027: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203856.51094: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203856.51114: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203856.51137: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203856.51232: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203856.51419: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203856.51422: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203856.95920: stdout chunk (state=3): >>> {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "hunspell-filesystem": [{"name": "hunspell-filesystem", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "google-noto-fonts-common": [{"name": "google-noto-fonts-common", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-mono-vf-fonts": [{"name": "google-noto-sans-mono-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-vf-fonts": [{"name": "google-noto-sans-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-serif-vf-fonts": [{"name": "google-noto-serif-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-mono-vf-fonts": [{"name": "redhat-mono-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-text-vf-fonts": [{"name": "redhat-text-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "default-fonts-core-sans": [{"name": "default-fonts-core-sans", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-fonts-en": [{"name": "langpacks-fonts-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-ucode-firmware": [{"name": "amd-ucode-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "atheros-firmware": [{"name": "atheros-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "brcmfmac-firmware": [{"name": "brcmfmac-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "cirrus-audio-firmware": [{"name": "cirrus-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-audio-firmware": [{"name": "intel-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "mt7xxx-firmware": [{"name": "mt7xxx-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nxpwireless-firmware": [{"name": "nxpwireless-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "realtek-firmware": [{"name": "realtek-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tiwilink-firmware": [{"name": "tiwilink-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-gpu-firmware": [{"name": "amd-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-gpu-firmware": [{"name": "intel-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nvidia-gpu-firmware": [{"name": "nvidia-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "xkeyboard-config": [{"name": "xkeyboard-config", "version": "2.41", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks"<<< 13731 1727203856.96076: stdout chunk (state=3): >>>: [{"name": "gawk-all-langpacks", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-data": [{"name": "vim-data", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20240107", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.379", "release": "10.1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.14.5", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.18", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "21.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.2.26", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-ng-compat": [{"name": "zlib-ng-compat", "version": "2.1.6", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.19", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.5", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.5", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.36", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.30", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.50", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.11", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.46.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.4", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "1.1", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "51.20230828cvs.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.6.2", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.1", "release": "9.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.17", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.9", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.10.0", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.5.0", "release": "1.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "hunspell-en-GB": [{"name": "hunspell-en-GB", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell-en-US": [{"name": "hunspell-en-US", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell": [{"name": "hunspell", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.6.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam-libs": [{"name": "pam-libs", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.69", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.15.0", "release": "3.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.19.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.80.4", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-libs": [{"name": "polkit-libs", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "101.2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.13", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.11.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.6.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.7", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.10", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.3", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.6", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.4", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.7.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.16", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.6.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.5", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "42.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.7", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sequoia": [{"name": "rpm-sequoia", "version": "1.6.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.29", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.79.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsecret": [{"name": "libsecret", "version": "0.21.2", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pinentry": [{"name": "pinentry", "version": "1.3.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libusb1": [{"name": "libusb1", "version": "1.0.27", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "4.0.4", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hunspell-en": [{"name": "hunspell-en", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.2.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm": [{"name": "gdbm", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libxkbcommon": [{"name": "libxkbcommon", "version": "1.7.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.23.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.7", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.10", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.8", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "35", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "661", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.6", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "memstrack": [{"name": "memstrack", "version": "0.2.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.25.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.15", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "duktape": [{"name": "duktape", "version": "2.7.0", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "22.el10.gating_test1", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse3-libs": [{"name": "fuse3-libs", "version": "3.16.2", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-envsubst": [{"name": "gettext-envsubst", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-runtime": [{"name": "gettext-runtime", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "58", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.1.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.11.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.14.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "28.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.62.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.7", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-gssapi": [{"name": "cyrus-sasl-gssapi", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11^20190603git9e74f2d", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "polkit": [{"name": "polkit", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-pkla-compat": [{"name": "polkit-pkla-compat", "version": "0.1", "release": "29.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "audit-rules": [{"name": "audit-rules", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20240905", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hasher": [{"name": "libkcapi-hasher", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.22.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "makedumpfile": [{"name": "makedumpfile", "version": "1.7.5", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.9.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.18.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss-fapi": [{"name": "tpm2-tss-fapi", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tools": [{"name": "tpm2-tools", "version": "5.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "76.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.81", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.15.0", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.32", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-krb5-common": [{"name": "sssd-krb5-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpdecimal": [{"name": "mpdecimal", "version": "2.5.1", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.3.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo-python-plugin": [{"name": "sudo-python-plugin", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.16.0", "release": "15.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.2", "release": "14.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "235", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.9", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.7.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.29", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kdump-utils": [{"name": "kdump-utils", "version": "1.0.43", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.13.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-libs": [{"name": "pcsc-lite-libs", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-ccid": [<<< 13731 1727203856.96199: stdout chunk (state=3): >>>{"name": "pcsc-lite-ccid", "version": "1.6.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite": [{"name": "pcsc-lite", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2-smime": [{"name": "gnupg2-smime", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.3", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.24", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.14.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "1.2304.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2408.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "8.el10", "epoch": 18, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.12.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.2.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.20", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwlwifi-dvm-firmware": [{"name": "iwlwifi-dvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwlwifi-mvm-firmware": [{"name": "iwlwifi-mvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "37.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.5", "release": "0.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.20", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.59", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.89", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.189", "release": "511.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.15", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.27", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2024.001", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20231213", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.61.000", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-locale": [{"name": "perl-locale", "version": "1.12", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.42", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.350", "release": "510.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.085", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.68", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "2.20", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.22", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.45", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.088", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.038", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.09", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "5.01", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.18", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.31", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.29", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.55", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.03", "release": "510.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.38", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.86", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.63", "release": "510.el10", "epoch": 5, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.32", "release": "510.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.37", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.241", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.58", "release": "2.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.54", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.78", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.91", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.56", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.17", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.21", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.44", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "23.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xxd": [{"name": "xxd", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.39", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "5.2.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "24.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.35", "release": "4.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.10", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.24.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.4.1", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.9.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.3.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpds-py": [{"name": "python3-rpds-py", "version": "0.17.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "23.2.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-referencing": [{"name": "python3-referencing", "version": "0.31.1", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.19", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema-specifications": [{"name": "python3-jsonschema-specifications", "version": "2023.11.2", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "4.19.1", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.5", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.2.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "2.1.3", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "3.1.4", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.3", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.33", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.9.0", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.8", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "69.0.3", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.5.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "6.0.1", "release": "18.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-charset-normalizer": [{"name": "python3-charset-normalizer", "version": "3.3.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.32.3", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dhcpcd": [{"name": "dhcpcd", "version": "10.0.6", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "24.1.4", "release": "17.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "device-mapper-event-libs": [{"name": "device-mapper-event-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "libaio": [{"name": "libaio", "version": "0.3.111", "release": "20.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-event": [{"name": "device-mapper-event", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "lvm2-libs": [{"name": "lvm2-libs", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper-persistent-data": [{"name": "device-mapper-persistent-data", "version": "1.0.11", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lvm2": [{"name": "lvm2", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.17", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.90", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} <<< 13731 1727203856.97958: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203856.97964: stdout chunk (state=3): >>><<< 13731 1727203856.97967: stderr chunk (state=3): >>><<< 13731 1727203856.98190: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "hunspell-filesystem": [{"name": "hunspell-filesystem", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "google-noto-fonts-common": [{"name": "google-noto-fonts-common", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-mono-vf-fonts": [{"name": "google-noto-sans-mono-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-vf-fonts": [{"name": "google-noto-sans-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-serif-vf-fonts": [{"name": "google-noto-serif-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-mono-vf-fonts": [{"name": "redhat-mono-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-text-vf-fonts": [{"name": "redhat-text-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "default-fonts-core-sans": [{"name": "default-fonts-core-sans", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-fonts-en": [{"name": "langpacks-fonts-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-ucode-firmware": [{"name": "amd-ucode-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "atheros-firmware": [{"name": "atheros-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "brcmfmac-firmware": [{"name": "brcmfmac-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "cirrus-audio-firmware": [{"name": "cirrus-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-audio-firmware": [{"name": "intel-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "mt7xxx-firmware": [{"name": "mt7xxx-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nxpwireless-firmware": [{"name": "nxpwireless-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "realtek-firmware": [{"name": "realtek-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tiwilink-firmware": [{"name": "tiwilink-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-gpu-firmware": [{"name": "amd-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-gpu-firmware": [{"name": "intel-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nvidia-gpu-firmware": [{"name": "nvidia-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "xkeyboard-config": [{"name": "xkeyboard-config", "version": "2.41", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-data": [{"name": "vim-data", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20240107", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.379", "release": "10.1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.14.5", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.18", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "21.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.2.26", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-ng-compat": [{"name": "zlib-ng-compat", "version": "2.1.6", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.19", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.5", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.5", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.36", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.30", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.50", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.11", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.46.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.4", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "1.1", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "51.20230828cvs.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.6.2", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.1", "release": "9.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.17", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.9", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.10.0", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.5.0", "release": "1.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "hunspell-en-GB": [{"name": "hunspell-en-GB", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell-en-US": [{"name": "hunspell-en-US", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell": [{"name": "hunspell", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.6.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam-libs": [{"name": "pam-libs", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.69", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.15.0", "release": "3.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.19.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.80.4", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-libs": [{"name": "polkit-libs", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "101.2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.13", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.11.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.6.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.7", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.10", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.3", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.6", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.4", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.7.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.16", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.6.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.5", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "42.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.7", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sequoia": [{"name": "rpm-sequoia", "version": "1.6.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.29", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.79.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsecret": [{"name": "libsecret", "version": "0.21.2", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pinentry": [{"name": "pinentry", "version": "1.3.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libusb1": [{"name": "libusb1", "version": "1.0.27", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "4.0.4", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hunspell-en": [{"name": "hunspell-en", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.2.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm": [{"name": "gdbm", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libxkbcommon": [{"name": "libxkbcommon", "version": "1.7.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.23.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.7", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.10", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.8", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "35", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "661", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.6", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "memstrack": [{"name": "memstrack", "version": "0.2.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.25.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.15", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "duktape": [{"name": "duktape", "version": "2.7.0", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "22.el10.gating_test1", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse3-libs": [{"name": "fuse3-libs", "version": "3.16.2", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-envsubst": [{"name": "gettext-envsubst", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-runtime": [{"name": "gettext-runtime", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "58", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.1.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.11.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.14.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "28.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.62.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.7", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-gssapi": [{"name": "cyrus-sasl-gssapi", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11^20190603git9e74f2d", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "polkit": [{"name": "polkit", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-pkla-compat": [{"name": "polkit-pkla-compat", "version": "0.1", "release": "29.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "audit-rules": [{"name": "audit-rules", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20240905", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hasher": [{"name": "libkcapi-hasher", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.22.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "makedumpfile": [{"name": "makedumpfile", "version": "1.7.5", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.9.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.18.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss-fapi": [{"name": "tpm2-tss-fapi", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tools": [{"name": "tpm2-tools", "version": "5.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "76.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.81", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.15.0", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.32", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-krb5-common": [{"name": "sssd-krb5-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpdecimal": [{"name": "mpdecimal", "version": "2.5.1", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.3.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo-python-plugin": [{"name": "sudo-python-plugin", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.16.0", "release": "15.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.2", "release": "14.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "235", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.9", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.7.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.29", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kdump-utils": [{"name": "kdump-utils", "version": "1.0.43", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.13.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-libs": [{"name": "pcsc-lite-libs", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-ccid": [{"name": "pcsc-lite-ccid", "version": "1.6.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite": [{"name": "pcsc-lite", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2-smime": [{"name": "gnupg2-smime", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.3", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.24", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.14.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "1.2304.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2408.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "8.el10", "epoch": 18, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.12.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.2.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.20", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwlwifi-dvm-firmware": [{"name": "iwlwifi-dvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwlwifi-mvm-firmware": [{"name": "iwlwifi-mvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "37.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.5", "release": "0.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.20", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.59", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.89", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.189", "release": "511.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.15", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.27", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2024.001", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20231213", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.61.000", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-locale": [{"name": "perl-locale", "version": "1.12", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.42", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.350", "release": "510.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.085", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.68", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "2.20", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.22", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.45", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.088", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.038", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.09", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "5.01", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.18", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.31", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.29", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.55", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.03", "release": "510.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.38", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.86", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.63", "release": "510.el10", "epoch": 5, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.32", "release": "510.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.37", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.241", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.58", "release": "2.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.54", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.78", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.91", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.56", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.17", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.21", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.44", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "23.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xxd": [{"name": "xxd", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.39", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "5.2.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "24.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.35", "release": "4.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.10", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.24.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.4.1", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.9.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.3.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpds-py": [{"name": "python3-rpds-py", "version": "0.17.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "23.2.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-referencing": [{"name": "python3-referencing", "version": "0.31.1", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.19", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema-specifications": [{"name": "python3-jsonschema-specifications", "version": "2023.11.2", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "4.19.1", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.5", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.2.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "2.1.3", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "3.1.4", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.3", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.33", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.9.0", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.8", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "69.0.3", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.5.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "6.0.1", "release": "18.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-charset-normalizer": [{"name": "python3-charset-normalizer", "version": "3.3.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.32.3", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dhcpcd": [{"name": "dhcpcd", "version": "10.0.6", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "24.1.4", "release": "17.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "device-mapper-event-libs": [{"name": "device-mapper-event-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "libaio": [{"name": "libaio", "version": "0.3.111", "release": "20.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-event": [{"name": "device-mapper-event", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "lvm2-libs": [{"name": "lvm2-libs", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper-persistent-data": [{"name": "device-mapper-persistent-data", "version": "1.0.11", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lvm2": [{"name": "lvm2", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.17", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.90", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203857.03411: done with _execute_module (package_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'package_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203857.03473: _low_level_execute_command(): starting 13731 1727203857.03479: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203856.3326702-16108-248345217445935/ > /dev/null 2>&1 && sleep 0' 13731 1727203857.04915: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203857.04922: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203857.05022: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203857.05025: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203857.05028: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203857.05030: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203857.05032: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203857.05033: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203857.05035: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203857.05037: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203857.05102: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203857.05364: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203857.05438: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203857.07225: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203857.07271: stderr chunk (state=3): >>><<< 13731 1727203857.07394: stdout chunk (state=3): >>><<< 13731 1727203857.07413: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203857.07419: handler run complete 13731 1727203857.09146: variable 'ansible_facts' from source: unknown 13731 1727203857.10345: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203857.14279: variable 'ansible_facts' from source: unknown 13731 1727203857.15172: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203857.16496: attempt loop complete, returning result 13731 1727203857.16500: _execute() done 13731 1727203857.16502: dumping result to json 13731 1727203857.16737: done dumping result, returning 13731 1727203857.16752: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which packages are installed [028d2410-947f-82dc-c122-0000000007cf] 13731 1727203857.16763: sending task result for task 028d2410-947f-82dc-c122-0000000007cf 13731 1727203857.21188: done sending task result for task 028d2410-947f-82dc-c122-0000000007cf 13731 1727203857.21192: WORKER PROCESS EXITING ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203857.21318: no more pending results, returning what we have 13731 1727203857.21322: results queue empty 13731 1727203857.21323: checking for any_errors_fatal 13731 1727203857.21327: done checking for any_errors_fatal 13731 1727203857.21328: checking for max_fail_percentage 13731 1727203857.21330: done checking for max_fail_percentage 13731 1727203857.21330: checking to see if all hosts have failed and the running result is not ok 13731 1727203857.21331: done checking to see if all hosts have failed 13731 1727203857.21332: getting the remaining hosts for this loop 13731 1727203857.21333: done getting the remaining hosts for this loop 13731 1727203857.21336: getting the next task for host managed-node3 13731 1727203857.21345: done getting next task for host managed-node3 13731 1727203857.21349: ^ task is: TASK: fedora.linux_system_roles.network : Print network provider 13731 1727203857.21354: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203857.21365: getting variables 13731 1727203857.21367: in VariableManager get_vars() 13731 1727203857.21428: Calling all_inventory to load vars for managed-node3 13731 1727203857.21431: Calling groups_inventory to load vars for managed-node3 13731 1727203857.21434: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203857.21443: Calling all_plugins_play to load vars for managed-node3 13731 1727203857.21446: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203857.21449: Calling groups_plugins_play to load vars for managed-node3 13731 1727203857.22941: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203857.25278: done with get_vars() 13731 1727203857.25318: done getting variables 13731 1727203857.25381: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Print network provider] ************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:7 Tuesday 24 September 2024 14:50:57 -0400 (0:00:01.008) 0:00:35.485 ***** 13731 1727203857.25537: entering _queue_task() for managed-node3/debug 13731 1727203857.26465: worker is 1 (out of 1 available) 13731 1727203857.26481: exiting _queue_task() for managed-node3/debug 13731 1727203857.26694: done queuing things up, now waiting for results queue to drain 13731 1727203857.26698: waiting for pending results... 13731 1727203857.27579: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Print network provider 13731 1727203857.28083: in run() - task 028d2410-947f-82dc-c122-000000000694 13731 1727203857.28091: variable 'ansible_search_path' from source: unknown 13731 1727203857.28096: variable 'ansible_search_path' from source: unknown 13731 1727203857.28100: calling self._execute() 13731 1727203857.28103: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203857.28106: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203857.28109: variable 'omit' from source: magic vars 13731 1727203857.28965: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.28989: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203857.29106: variable 'omit' from source: magic vars 13731 1727203857.29114: variable 'omit' from source: magic vars 13731 1727203857.29328: variable 'network_provider' from source: set_fact 13731 1727203857.29803: variable 'omit' from source: magic vars 13731 1727203857.29807: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203857.29810: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203857.30102: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203857.30105: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203857.30108: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203857.30110: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203857.30113: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203857.30116: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203857.30893: Set connection var ansible_pipelining to False 13731 1727203857.30896: Set connection var ansible_shell_type to sh 13731 1727203857.30899: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203857.30902: Set connection var ansible_connection to ssh 13731 1727203857.30904: Set connection var ansible_shell_executable to /bin/sh 13731 1727203857.30906: Set connection var ansible_timeout to 10 13731 1727203857.30908: variable 'ansible_shell_executable' from source: unknown 13731 1727203857.30910: variable 'ansible_connection' from source: unknown 13731 1727203857.30913: variable 'ansible_module_compression' from source: unknown 13731 1727203857.30915: variable 'ansible_shell_type' from source: unknown 13731 1727203857.30916: variable 'ansible_shell_executable' from source: unknown 13731 1727203857.30919: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203857.30921: variable 'ansible_pipelining' from source: unknown 13731 1727203857.31005: variable 'ansible_timeout' from source: unknown 13731 1727203857.31015: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203857.31439: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203857.31677: variable 'omit' from source: magic vars 13731 1727203857.31690: starting attempt loop 13731 1727203857.31697: running the handler 13731 1727203857.31745: handler run complete 13731 1727203857.31784: attempt loop complete, returning result 13731 1727203857.31882: _execute() done 13731 1727203857.31889: dumping result to json 13731 1727203857.31897: done dumping result, returning 13731 1727203857.31909: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Print network provider [028d2410-947f-82dc-c122-000000000694] 13731 1727203857.31923: sending task result for task 028d2410-947f-82dc-c122-000000000694 13731 1727203857.32281: done sending task result for task 028d2410-947f-82dc-c122-000000000694 13731 1727203857.32285: WORKER PROCESS EXITING ok: [managed-node3] => {} MSG: Using network provider: nm 13731 1727203857.32349: no more pending results, returning what we have 13731 1727203857.32353: results queue empty 13731 1727203857.32354: checking for any_errors_fatal 13731 1727203857.32366: done checking for any_errors_fatal 13731 1727203857.32367: checking for max_fail_percentage 13731 1727203857.32368: done checking for max_fail_percentage 13731 1727203857.32369: checking to see if all hosts have failed and the running result is not ok 13731 1727203857.32370: done checking to see if all hosts have failed 13731 1727203857.32371: getting the remaining hosts for this loop 13731 1727203857.32373: done getting the remaining hosts for this loop 13731 1727203857.32378: getting the next task for host managed-node3 13731 1727203857.32387: done getting next task for host managed-node3 13731 1727203857.32392: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13731 1727203857.32397: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203857.32410: getting variables 13731 1727203857.32412: in VariableManager get_vars() 13731 1727203857.32450: Calling all_inventory to load vars for managed-node3 13731 1727203857.32453: Calling groups_inventory to load vars for managed-node3 13731 1727203857.32455: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203857.32467: Calling all_plugins_play to load vars for managed-node3 13731 1727203857.32470: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203857.32473: Calling groups_plugins_play to load vars for managed-node3 13731 1727203857.35781: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203857.38313: done with get_vars() 13731 1727203857.38345: done getting variables 13731 1727203857.38517: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:11 Tuesday 24 September 2024 14:50:57 -0400 (0:00:00.131) 0:00:35.616 ***** 13731 1727203857.38562: entering _queue_task() for managed-node3/fail 13731 1727203857.39316: worker is 1 (out of 1 available) 13731 1727203857.39331: exiting _queue_task() for managed-node3/fail 13731 1727203857.39346: done queuing things up, now waiting for results queue to drain 13731 1727203857.39347: waiting for pending results... 13731 1727203857.39811: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13731 1727203857.39996: in run() - task 028d2410-947f-82dc-c122-000000000695 13731 1727203857.40021: variable 'ansible_search_path' from source: unknown 13731 1727203857.40071: variable 'ansible_search_path' from source: unknown 13731 1727203857.40077: calling self._execute() 13731 1727203857.40182: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203857.40195: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203857.40209: variable 'omit' from source: magic vars 13731 1727203857.40637: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.40660: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203857.40828: variable 'network_state' from source: role '' defaults 13731 1727203857.40835: Evaluated conditional (network_state != {}): False 13731 1727203857.40838: when evaluation is False, skipping this task 13731 1727203857.40841: _execute() done 13731 1727203857.40844: dumping result to json 13731 1727203857.40846: done dumping result, returning 13731 1727203857.40852: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider [028d2410-947f-82dc-c122-000000000695] 13731 1727203857.40873: sending task result for task 028d2410-947f-82dc-c122-000000000695 13731 1727203857.41074: done sending task result for task 028d2410-947f-82dc-c122-000000000695 13731 1727203857.41079: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203857.41140: no more pending results, returning what we have 13731 1727203857.41146: results queue empty 13731 1727203857.41147: checking for any_errors_fatal 13731 1727203857.41154: done checking for any_errors_fatal 13731 1727203857.41155: checking for max_fail_percentage 13731 1727203857.41158: done checking for max_fail_percentage 13731 1727203857.41159: checking to see if all hosts have failed and the running result is not ok 13731 1727203857.41159: done checking to see if all hosts have failed 13731 1727203857.41160: getting the remaining hosts for this loop 13731 1727203857.41162: done getting the remaining hosts for this loop 13731 1727203857.41166: getting the next task for host managed-node3 13731 1727203857.41177: done getting next task for host managed-node3 13731 1727203857.41181: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13731 1727203857.41302: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203857.41327: getting variables 13731 1727203857.41329: in VariableManager get_vars() 13731 1727203857.41366: Calling all_inventory to load vars for managed-node3 13731 1727203857.41369: Calling groups_inventory to load vars for managed-node3 13731 1727203857.41371: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203857.41420: Calling all_plugins_play to load vars for managed-node3 13731 1727203857.41424: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203857.41427: Calling groups_plugins_play to load vars for managed-node3 13731 1727203857.43916: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203857.47079: done with get_vars() 13731 1727203857.47118: done getting variables 13731 1727203857.47232: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:18 Tuesday 24 September 2024 14:50:57 -0400 (0:00:00.087) 0:00:35.704 ***** 13731 1727203857.47363: entering _queue_task() for managed-node3/fail 13731 1727203857.47877: worker is 1 (out of 1 available) 13731 1727203857.47890: exiting _queue_task() for managed-node3/fail 13731 1727203857.47900: done queuing things up, now waiting for results queue to drain 13731 1727203857.47902: waiting for pending results... 13731 1727203857.48202: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13731 1727203857.48381: in run() - task 028d2410-947f-82dc-c122-000000000696 13731 1727203857.48384: variable 'ansible_search_path' from source: unknown 13731 1727203857.48386: variable 'ansible_search_path' from source: unknown 13731 1727203857.48389: calling self._execute() 13731 1727203857.48463: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203857.48473: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203857.48488: variable 'omit' from source: magic vars 13731 1727203857.48916: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.48932: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203857.49079: variable 'network_state' from source: role '' defaults 13731 1727203857.49095: Evaluated conditional (network_state != {}): False 13731 1727203857.49102: when evaluation is False, skipping this task 13731 1727203857.49109: _execute() done 13731 1727203857.49115: dumping result to json 13731 1727203857.49121: done dumping result, returning 13731 1727203857.49131: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 [028d2410-947f-82dc-c122-000000000696] 13731 1727203857.49141: sending task result for task 028d2410-947f-82dc-c122-000000000696 13731 1727203857.49343: done sending task result for task 028d2410-947f-82dc-c122-000000000696 13731 1727203857.49347: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203857.49424: no more pending results, returning what we have 13731 1727203857.49429: results queue empty 13731 1727203857.49429: checking for any_errors_fatal 13731 1727203857.49438: done checking for any_errors_fatal 13731 1727203857.49438: checking for max_fail_percentage 13731 1727203857.49441: done checking for max_fail_percentage 13731 1727203857.49442: checking to see if all hosts have failed and the running result is not ok 13731 1727203857.49443: done checking to see if all hosts have failed 13731 1727203857.49444: getting the remaining hosts for this loop 13731 1727203857.49446: done getting the remaining hosts for this loop 13731 1727203857.49450: getting the next task for host managed-node3 13731 1727203857.49460: done getting next task for host managed-node3 13731 1727203857.49464: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13731 1727203857.49469: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203857.49495: getting variables 13731 1727203857.49497: in VariableManager get_vars() 13731 1727203857.49537: Calling all_inventory to load vars for managed-node3 13731 1727203857.49540: Calling groups_inventory to load vars for managed-node3 13731 1727203857.49543: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203857.49555: Calling all_plugins_play to load vars for managed-node3 13731 1727203857.49558: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203857.49561: Calling groups_plugins_play to load vars for managed-node3 13731 1727203857.53401: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203857.55434: done with get_vars() 13731 1727203857.55480: done getting variables 13731 1727203857.55717: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:25 Tuesday 24 September 2024 14:50:57 -0400 (0:00:00.083) 0:00:35.788 ***** 13731 1727203857.55871: entering _queue_task() for managed-node3/fail 13731 1727203857.56279: worker is 1 (out of 1 available) 13731 1727203857.56292: exiting _queue_task() for managed-node3/fail 13731 1727203857.56491: done queuing things up, now waiting for results queue to drain 13731 1727203857.56493: waiting for pending results... 13731 1727203857.57016: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13731 1727203857.57215: in run() - task 028d2410-947f-82dc-c122-000000000697 13731 1727203857.57349: variable 'ansible_search_path' from source: unknown 13731 1727203857.57358: variable 'ansible_search_path' from source: unknown 13731 1727203857.57442: calling self._execute() 13731 1727203857.57795: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203857.57801: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203857.57920: variable 'omit' from source: magic vars 13731 1727203857.58616: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.58637: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203857.58996: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203857.63100: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203857.63252: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203857.63421: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203857.63424: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203857.63509: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203857.63624: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.63673: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.63705: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.63836: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.63892: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.64152: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.64209: Evaluated conditional (ansible_distribution_major_version | int > 9): True 13731 1727203857.64341: variable 'ansible_distribution' from source: facts 13731 1727203857.64350: variable '__network_rh_distros' from source: role '' defaults 13731 1727203857.64364: Evaluated conditional (ansible_distribution in __network_rh_distros): True 13731 1727203857.64911: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.64946: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.65002: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.65048: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.65122: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.65204: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.65233: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.65262: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.65438: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.65441: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.65468: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.65498: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.65535: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.65583: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.65602: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.66066: variable 'network_connections' from source: task vars 13731 1727203857.66071: variable 'port2_profile' from source: play vars 13731 1727203857.66073: variable 'port2_profile' from source: play vars 13731 1727203857.66093: variable 'port1_profile' from source: play vars 13731 1727203857.66157: variable 'port1_profile' from source: play vars 13731 1727203857.66170: variable 'controller_profile' from source: play vars 13731 1727203857.66243: variable 'controller_profile' from source: play vars 13731 1727203857.66257: variable 'network_state' from source: role '' defaults 13731 1727203857.66340: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203857.66741: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203857.66744: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203857.66746: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203857.66793: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203857.66907: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203857.66933: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203857.66992: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.67030: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203857.67081: Evaluated conditional (network_connections | selectattr("type", "defined") | selectattr("type", "match", "^team$") | list | length > 0 or network_state.get("interfaces", []) | selectattr("type", "defined") | selectattr("type", "match", "^team$") | list | length > 0): False 13731 1727203857.67085: when evaluation is False, skipping this task 13731 1727203857.67087: _execute() done 13731 1727203857.67189: dumping result to json 13731 1727203857.67192: done dumping result, returning 13731 1727203857.67194: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later [028d2410-947f-82dc-c122-000000000697] 13731 1727203857.67197: sending task result for task 028d2410-947f-82dc-c122-000000000697 13731 1727203857.67277: done sending task result for task 028d2410-947f-82dc-c122-000000000697 13731 1727203857.67281: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_connections | selectattr(\"type\", \"defined\") | selectattr(\"type\", \"match\", \"^team$\") | list | length > 0 or network_state.get(\"interfaces\", []) | selectattr(\"type\", \"defined\") | selectattr(\"type\", \"match\", \"^team$\") | list | length > 0", "skip_reason": "Conditional result was False" } 13731 1727203857.67343: no more pending results, returning what we have 13731 1727203857.67351: results queue empty 13731 1727203857.67352: checking for any_errors_fatal 13731 1727203857.67360: done checking for any_errors_fatal 13731 1727203857.67361: checking for max_fail_percentage 13731 1727203857.67363: done checking for max_fail_percentage 13731 1727203857.67364: checking to see if all hosts have failed and the running result is not ok 13731 1727203857.67365: done checking to see if all hosts have failed 13731 1727203857.67366: getting the remaining hosts for this loop 13731 1727203857.67368: done getting the remaining hosts for this loop 13731 1727203857.67371: getting the next task for host managed-node3 13731 1727203857.67382: done getting next task for host managed-node3 13731 1727203857.67387: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13731 1727203857.67394: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203857.67412: getting variables 13731 1727203857.67415: in VariableManager get_vars() 13731 1727203857.67458: Calling all_inventory to load vars for managed-node3 13731 1727203857.67461: Calling groups_inventory to load vars for managed-node3 13731 1727203857.67464: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203857.67657: Calling all_plugins_play to load vars for managed-node3 13731 1727203857.67663: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203857.67667: Calling groups_plugins_play to load vars for managed-node3 13731 1727203857.70277: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203857.72590: done with get_vars() 13731 1727203857.72616: done getting variables 13731 1727203857.72689: Loading ActionModule 'dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:36 Tuesday 24 September 2024 14:50:57 -0400 (0:00:00.169) 0:00:35.958 ***** 13731 1727203857.72723: entering _queue_task() for managed-node3/dnf 13731 1727203857.73285: worker is 1 (out of 1 available) 13731 1727203857.73296: exiting _queue_task() for managed-node3/dnf 13731 1727203857.73307: done queuing things up, now waiting for results queue to drain 13731 1727203857.73309: waiting for pending results... 13731 1727203857.73660: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13731 1727203857.73665: in run() - task 028d2410-947f-82dc-c122-000000000698 13731 1727203857.73668: variable 'ansible_search_path' from source: unknown 13731 1727203857.73671: variable 'ansible_search_path' from source: unknown 13731 1727203857.73702: calling self._execute() 13731 1727203857.73806: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203857.73817: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203857.73830: variable 'omit' from source: magic vars 13731 1727203857.74222: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.74239: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203857.74456: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203857.76810: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203857.76981: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203857.76986: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203857.76988: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203857.77002: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203857.77083: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.77155: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.77190: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.77247: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.77268: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.77394: variable 'ansible_distribution' from source: facts 13731 1727203857.77405: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.77428: Evaluated conditional (ansible_distribution == 'Fedora' or ansible_distribution_major_version | int > 7): True 13731 1727203857.77572: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203857.77706: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.77761: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.77766: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.77818: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.77839: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.77980: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.77984: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.77986: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.78005: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.78026: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.78069: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.78110: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.78142: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.78187: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.78216: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.78414: variable 'network_connections' from source: task vars 13731 1727203857.78418: variable 'port2_profile' from source: play vars 13731 1727203857.78482: variable 'port2_profile' from source: play vars 13731 1727203857.78498: variable 'port1_profile' from source: play vars 13731 1727203857.78572: variable 'port1_profile' from source: play vars 13731 1727203857.78632: variable 'controller_profile' from source: play vars 13731 1727203857.78661: variable 'controller_profile' from source: play vars 13731 1727203857.78742: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203857.78930: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203857.78989: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203857.79024: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203857.79057: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203857.79179: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203857.79197: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203857.79369: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.79373: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203857.79378: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203857.79634: variable 'network_connections' from source: task vars 13731 1727203857.79645: variable 'port2_profile' from source: play vars 13731 1727203857.79854: variable 'port2_profile' from source: play vars 13731 1727203857.79868: variable 'port1_profile' from source: play vars 13731 1727203857.79939: variable 'port1_profile' from source: play vars 13731 1727203857.79952: variable 'controller_profile' from source: play vars 13731 1727203857.80027: variable 'controller_profile' from source: play vars 13731 1727203857.80056: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203857.80064: when evaluation is False, skipping this task 13731 1727203857.80070: _execute() done 13731 1727203857.80079: dumping result to json 13731 1727203857.80087: done dumping result, returning 13731 1727203857.80099: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000698] 13731 1727203857.80108: sending task result for task 028d2410-947f-82dc-c122-000000000698 skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203857.80431: no more pending results, returning what we have 13731 1727203857.80435: results queue empty 13731 1727203857.80436: checking for any_errors_fatal 13731 1727203857.80444: done checking for any_errors_fatal 13731 1727203857.80445: checking for max_fail_percentage 13731 1727203857.80447: done checking for max_fail_percentage 13731 1727203857.80448: checking to see if all hosts have failed and the running result is not ok 13731 1727203857.80449: done checking to see if all hosts have failed 13731 1727203857.80449: getting the remaining hosts for this loop 13731 1727203857.80451: done getting the remaining hosts for this loop 13731 1727203857.80454: getting the next task for host managed-node3 13731 1727203857.80463: done getting next task for host managed-node3 13731 1727203857.80467: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13731 1727203857.80473: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203857.80498: getting variables 13731 1727203857.80500: in VariableManager get_vars() 13731 1727203857.80541: Calling all_inventory to load vars for managed-node3 13731 1727203857.80544: Calling groups_inventory to load vars for managed-node3 13731 1727203857.80547: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203857.80559: Calling all_plugins_play to load vars for managed-node3 13731 1727203857.80563: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203857.80566: Calling groups_plugins_play to load vars for managed-node3 13731 1727203857.81092: done sending task result for task 028d2410-947f-82dc-c122-000000000698 13731 1727203857.81095: WORKER PROCESS EXITING 13731 1727203857.82224: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203857.83858: done with get_vars() 13731 1727203857.83924: done getting variables redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13731 1727203857.84259: Loading ActionModule 'ansible_collections.ansible.builtin.plugins.action.dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:48 Tuesday 24 September 2024 14:50:57 -0400 (0:00:00.116) 0:00:36.074 ***** 13731 1727203857.84401: entering _queue_task() for managed-node3/yum 13731 1727203857.85072: worker is 1 (out of 1 available) 13731 1727203857.85145: exiting _queue_task() for managed-node3/yum 13731 1727203857.85157: done queuing things up, now waiting for results queue to drain 13731 1727203857.85159: waiting for pending results... 13731 1727203857.85913: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13731 1727203857.86457: in run() - task 028d2410-947f-82dc-c122-000000000699 13731 1727203857.86477: variable 'ansible_search_path' from source: unknown 13731 1727203857.86483: variable 'ansible_search_path' from source: unknown 13731 1727203857.86521: calling self._execute() 13731 1727203857.87110: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203857.87115: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203857.87118: variable 'omit' from source: magic vars 13731 1727203857.87732: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.87745: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203857.88255: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203857.91281: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203857.91358: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203857.91399: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203857.91433: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203857.91466: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203857.91695: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203857.91698: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203857.91711: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203857.91755: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203857.91772: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203857.91946: variable 'ansible_distribution_major_version' from source: facts 13731 1727203857.91949: Evaluated conditional (ansible_distribution_major_version | int < 8): False 13731 1727203857.91951: when evaluation is False, skipping this task 13731 1727203857.91953: _execute() done 13731 1727203857.91956: dumping result to json 13731 1727203857.91958: done dumping result, returning 13731 1727203857.91960: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000699] 13731 1727203857.91963: sending task result for task 028d2410-947f-82dc-c122-000000000699 13731 1727203857.92341: done sending task result for task 028d2410-947f-82dc-c122-000000000699 13731 1727203857.92349: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version | int < 8", "skip_reason": "Conditional result was False" } 13731 1727203857.92405: no more pending results, returning what we have 13731 1727203857.92410: results queue empty 13731 1727203857.92411: checking for any_errors_fatal 13731 1727203857.92417: done checking for any_errors_fatal 13731 1727203857.92418: checking for max_fail_percentage 13731 1727203857.92420: done checking for max_fail_percentage 13731 1727203857.92421: checking to see if all hosts have failed and the running result is not ok 13731 1727203857.92421: done checking to see if all hosts have failed 13731 1727203857.92422: getting the remaining hosts for this loop 13731 1727203857.92424: done getting the remaining hosts for this loop 13731 1727203857.92428: getting the next task for host managed-node3 13731 1727203857.92435: done getting next task for host managed-node3 13731 1727203857.92439: ^ task is: TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13731 1727203857.92444: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203857.92515: getting variables 13731 1727203857.92517: in VariableManager get_vars() 13731 1727203857.92553: Calling all_inventory to load vars for managed-node3 13731 1727203857.92556: Calling groups_inventory to load vars for managed-node3 13731 1727203857.92559: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203857.92622: Calling all_plugins_play to load vars for managed-node3 13731 1727203857.92626: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203857.92630: Calling groups_plugins_play to load vars for managed-node3 13731 1727203857.95297: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203858.03261: done with get_vars() 13731 1727203858.03297: done getting variables 13731 1727203858.03347: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:60 Tuesday 24 September 2024 14:50:58 -0400 (0:00:00.189) 0:00:36.264 ***** 13731 1727203858.03388: entering _queue_task() for managed-node3/fail 13731 1727203858.03971: worker is 1 (out of 1 available) 13731 1727203858.03984: exiting _queue_task() for managed-node3/fail 13731 1727203858.03994: done queuing things up, now waiting for results queue to drain 13731 1727203858.03996: waiting for pending results... 13731 1727203858.04229: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13731 1727203858.04467: in run() - task 028d2410-947f-82dc-c122-00000000069a 13731 1727203858.04483: variable 'ansible_search_path' from source: unknown 13731 1727203858.04492: variable 'ansible_search_path' from source: unknown 13731 1727203858.04622: calling self._execute() 13731 1727203858.04642: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203858.04645: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203858.04649: variable 'omit' from source: magic vars 13731 1727203858.05198: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.05215: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203858.05412: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203858.05732: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203858.08625: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203858.08747: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203858.08783: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203858.08827: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203858.08852: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203858.09003: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.09008: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.09011: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.09059: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.09072: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.09119: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.09145: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.09195: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.09295: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.09298: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.09300: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.09317: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.09339: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.09390: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.09414: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.09758: variable 'network_connections' from source: task vars 13731 1727203858.09771: variable 'port2_profile' from source: play vars 13731 1727203858.09854: variable 'port2_profile' from source: play vars 13731 1727203858.09864: variable 'port1_profile' from source: play vars 13731 1727203858.09992: variable 'port1_profile' from source: play vars 13731 1727203858.09995: variable 'controller_profile' from source: play vars 13731 1727203858.10094: variable 'controller_profile' from source: play vars 13731 1727203858.10154: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203858.10667: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203858.10714: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203858.10742: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203858.10778: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203858.11054: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203858.11060: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203858.11063: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.11246: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203858.11436: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203858.12100: variable 'network_connections' from source: task vars 13731 1727203858.12104: variable 'port2_profile' from source: play vars 13731 1727203858.12179: variable 'port2_profile' from source: play vars 13731 1727203858.12190: variable 'port1_profile' from source: play vars 13731 1727203858.12256: variable 'port1_profile' from source: play vars 13731 1727203858.12264: variable 'controller_profile' from source: play vars 13731 1727203858.12327: variable 'controller_profile' from source: play vars 13731 1727203858.12418: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203858.12429: when evaluation is False, skipping this task 13731 1727203858.12432: _execute() done 13731 1727203858.12434: dumping result to json 13731 1727203858.12436: done dumping result, returning 13731 1727203858.12438: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces [028d2410-947f-82dc-c122-00000000069a] 13731 1727203858.12440: sending task result for task 028d2410-947f-82dc-c122-00000000069a 13731 1727203858.12745: done sending task result for task 028d2410-947f-82dc-c122-00000000069a 13731 1727203858.12749: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203858.12799: no more pending results, returning what we have 13731 1727203858.12802: results queue empty 13731 1727203858.12803: checking for any_errors_fatal 13731 1727203858.12809: done checking for any_errors_fatal 13731 1727203858.12810: checking for max_fail_percentage 13731 1727203858.12812: done checking for max_fail_percentage 13731 1727203858.12812: checking to see if all hosts have failed and the running result is not ok 13731 1727203858.12813: done checking to see if all hosts have failed 13731 1727203858.12814: getting the remaining hosts for this loop 13731 1727203858.12816: done getting the remaining hosts for this loop 13731 1727203858.12819: getting the next task for host managed-node3 13731 1727203858.12826: done getting next task for host managed-node3 13731 1727203858.12830: ^ task is: TASK: fedora.linux_system_roles.network : Install packages 13731 1727203858.12836: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203858.12852: getting variables 13731 1727203858.12854: in VariableManager get_vars() 13731 1727203858.13033: Calling all_inventory to load vars for managed-node3 13731 1727203858.13037: Calling groups_inventory to load vars for managed-node3 13731 1727203858.13039: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203858.13048: Calling all_plugins_play to load vars for managed-node3 13731 1727203858.13051: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203858.13055: Calling groups_plugins_play to load vars for managed-node3 13731 1727203858.15434: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203858.18641: done with get_vars() 13731 1727203858.18681: done getting variables 13731 1727203858.18752: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install packages] ******************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:73 Tuesday 24 September 2024 14:50:58 -0400 (0:00:00.154) 0:00:36.419 ***** 13731 1727203858.18819: entering _queue_task() for managed-node3/package 13731 1727203858.19233: worker is 1 (out of 1 available) 13731 1727203858.19247: exiting _queue_task() for managed-node3/package 13731 1727203858.19264: done queuing things up, now waiting for results queue to drain 13731 1727203858.19266: waiting for pending results... 13731 1727203858.19635: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install packages 13731 1727203858.19748: in run() - task 028d2410-947f-82dc-c122-00000000069b 13731 1727203858.19757: variable 'ansible_search_path' from source: unknown 13731 1727203858.19763: variable 'ansible_search_path' from source: unknown 13731 1727203858.19836: calling self._execute() 13731 1727203858.19874: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203858.19943: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203858.19951: variable 'omit' from source: magic vars 13731 1727203858.20486: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.20490: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203858.20592: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203858.20932: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203858.20996: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203858.21072: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203858.21169: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203858.21304: variable 'network_packages' from source: role '' defaults 13731 1727203858.21414: variable '__network_provider_setup' from source: role '' defaults 13731 1727203858.21435: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203858.21503: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203858.21513: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203858.21599: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203858.21834: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203858.25972: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203858.26120: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203858.26264: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203858.26299: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203858.26325: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203858.26591: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.26626: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.26651: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.26748: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.26764: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.26954: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.26969: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.27179: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.27201: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.27213: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.27882: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13731 1727203858.27885: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.27904: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.27934: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.27970: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.27987: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.28080: variable 'ansible_python' from source: facts 13731 1727203858.28215: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13731 1727203858.28218: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203858.28291: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203858.28491: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.28515: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.28539: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.28812: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.28816: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.28818: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.28829: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.28831: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.28834: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.28836: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.28944: variable 'network_connections' from source: task vars 13731 1727203858.28983: variable 'port2_profile' from source: play vars 13731 1727203858.29053: variable 'port2_profile' from source: play vars 13731 1727203858.29065: variable 'port1_profile' from source: play vars 13731 1727203858.29415: variable 'port1_profile' from source: play vars 13731 1727203858.29425: variable 'controller_profile' from source: play vars 13731 1727203858.29524: variable 'controller_profile' from source: play vars 13731 1727203858.29936: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203858.29963: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203858.30129: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.30132: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203858.30391: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203858.30931: variable 'network_connections' from source: task vars 13731 1727203858.30938: variable 'port2_profile' from source: play vars 13731 1727203858.31314: variable 'port2_profile' from source: play vars 13731 1727203858.31317: variable 'port1_profile' from source: play vars 13731 1727203858.31511: variable 'port1_profile' from source: play vars 13731 1727203858.31514: variable 'controller_profile' from source: play vars 13731 1727203858.31832: variable 'controller_profile' from source: play vars 13731 1727203858.31836: variable '__network_packages_default_wireless' from source: role '' defaults 13731 1727203858.31923: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203858.32756: variable 'network_connections' from source: task vars 13731 1727203858.32762: variable 'port2_profile' from source: play vars 13731 1727203858.32938: variable 'port2_profile' from source: play vars 13731 1727203858.32945: variable 'port1_profile' from source: play vars 13731 1727203858.33009: variable 'port1_profile' from source: play vars 13731 1727203858.33016: variable 'controller_profile' from source: play vars 13731 1727203858.33255: variable 'controller_profile' from source: play vars 13731 1727203858.33292: variable '__network_packages_default_team' from source: role '' defaults 13731 1727203858.33430: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203858.33877: variable 'network_connections' from source: task vars 13731 1727203858.33900: variable 'port2_profile' from source: play vars 13731 1727203858.33955: variable 'port2_profile' from source: play vars 13731 1727203858.33965: variable 'port1_profile' from source: play vars 13731 1727203858.34028: variable 'port1_profile' from source: play vars 13731 1727203858.34034: variable 'controller_profile' from source: play vars 13731 1727203858.34097: variable 'controller_profile' from source: play vars 13731 1727203858.34162: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203858.34226: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203858.34229: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203858.34285: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203858.34663: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13731 1727203858.35354: variable 'network_connections' from source: task vars 13731 1727203858.35358: variable 'port2_profile' from source: play vars 13731 1727203858.35635: variable 'port2_profile' from source: play vars 13731 1727203858.35641: variable 'port1_profile' from source: play vars 13731 1727203858.35699: variable 'port1_profile' from source: play vars 13731 1727203858.35707: variable 'controller_profile' from source: play vars 13731 1727203858.35824: variable 'controller_profile' from source: play vars 13731 1727203858.35874: variable 'ansible_distribution' from source: facts 13731 1727203858.35881: variable '__network_rh_distros' from source: role '' defaults 13731 1727203858.35888: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.35904: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13731 1727203858.36132: variable 'ansible_distribution' from source: facts 13731 1727203858.36136: variable '__network_rh_distros' from source: role '' defaults 13731 1727203858.36141: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.36157: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13731 1727203858.36345: variable 'ansible_distribution' from source: facts 13731 1727203858.36349: variable '__network_rh_distros' from source: role '' defaults 13731 1727203858.36486: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.36520: variable 'network_provider' from source: set_fact 13731 1727203858.36533: variable 'ansible_facts' from source: unknown 13731 1727203858.38698: Evaluated conditional (not network_packages is subset(ansible_facts.packages.keys())): False 13731 1727203858.38700: when evaluation is False, skipping this task 13731 1727203858.38702: _execute() done 13731 1727203858.38704: dumping result to json 13731 1727203858.38706: done dumping result, returning 13731 1727203858.38708: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install packages [028d2410-947f-82dc-c122-00000000069b] 13731 1727203858.38709: sending task result for task 028d2410-947f-82dc-c122-00000000069b 13731 1727203858.38778: done sending task result for task 028d2410-947f-82dc-c122-00000000069b 13731 1727203858.38782: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "not network_packages is subset(ansible_facts.packages.keys())", "skip_reason": "Conditional result was False" } 13731 1727203858.38840: no more pending results, returning what we have 13731 1727203858.38845: results queue empty 13731 1727203858.38845: checking for any_errors_fatal 13731 1727203858.38853: done checking for any_errors_fatal 13731 1727203858.38854: checking for max_fail_percentage 13731 1727203858.38856: done checking for max_fail_percentage 13731 1727203858.38857: checking to see if all hosts have failed and the running result is not ok 13731 1727203858.38858: done checking to see if all hosts have failed 13731 1727203858.38859: getting the remaining hosts for this loop 13731 1727203858.38861: done getting the remaining hosts for this loop 13731 1727203858.38869: getting the next task for host managed-node3 13731 1727203858.38880: done getting next task for host managed-node3 13731 1727203858.38885: ^ task is: TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13731 1727203858.38890: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203858.38910: getting variables 13731 1727203858.38912: in VariableManager get_vars() 13731 1727203858.38951: Calling all_inventory to load vars for managed-node3 13731 1727203858.38954: Calling groups_inventory to load vars for managed-node3 13731 1727203858.38957: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203858.38967: Calling all_plugins_play to load vars for managed-node3 13731 1727203858.38971: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203858.38974: Calling groups_plugins_play to load vars for managed-node3 13731 1727203858.42111: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203858.46202: done with get_vars() 13731 1727203858.46403: done getting variables 13731 1727203858.46467: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:85 Tuesday 24 September 2024 14:50:58 -0400 (0:00:00.276) 0:00:36.696 ***** 13731 1727203858.46507: entering _queue_task() for managed-node3/package 13731 1727203858.47254: worker is 1 (out of 1 available) 13731 1727203858.47269: exiting _queue_task() for managed-node3/package 13731 1727203858.47284: done queuing things up, now waiting for results queue to drain 13731 1727203858.47286: waiting for pending results... 13731 1727203858.47995: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13731 1727203858.48170: in run() - task 028d2410-947f-82dc-c122-00000000069c 13731 1727203858.48263: variable 'ansible_search_path' from source: unknown 13731 1727203858.48312: variable 'ansible_search_path' from source: unknown 13731 1727203858.48415: calling self._execute() 13731 1727203858.48746: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203858.48750: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203858.48753: variable 'omit' from source: magic vars 13731 1727203858.49766: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.49843: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203858.50192: variable 'network_state' from source: role '' defaults 13731 1727203858.50301: Evaluated conditional (network_state != {}): False 13731 1727203858.50306: when evaluation is False, skipping this task 13731 1727203858.50317: _execute() done 13731 1727203858.50322: dumping result to json 13731 1727203858.50325: done dumping result, returning 13731 1727203858.50384: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable [028d2410-947f-82dc-c122-00000000069c] 13731 1727203858.50389: sending task result for task 028d2410-947f-82dc-c122-00000000069c 13731 1727203858.50669: done sending task result for task 028d2410-947f-82dc-c122-00000000069c 13731 1727203858.50673: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203858.50724: no more pending results, returning what we have 13731 1727203858.50729: results queue empty 13731 1727203858.50729: checking for any_errors_fatal 13731 1727203858.50736: done checking for any_errors_fatal 13731 1727203858.50736: checking for max_fail_percentage 13731 1727203858.50738: done checking for max_fail_percentage 13731 1727203858.50739: checking to see if all hosts have failed and the running result is not ok 13731 1727203858.50739: done checking to see if all hosts have failed 13731 1727203858.50740: getting the remaining hosts for this loop 13731 1727203858.50742: done getting the remaining hosts for this loop 13731 1727203858.50745: getting the next task for host managed-node3 13731 1727203858.50755: done getting next task for host managed-node3 13731 1727203858.50759: ^ task is: TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13731 1727203858.50765: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203858.50786: getting variables 13731 1727203858.50788: in VariableManager get_vars() 13731 1727203858.50829: Calling all_inventory to load vars for managed-node3 13731 1727203858.50832: Calling groups_inventory to load vars for managed-node3 13731 1727203858.50835: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203858.50848: Calling all_plugins_play to load vars for managed-node3 13731 1727203858.50851: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203858.50854: Calling groups_plugins_play to load vars for managed-node3 13731 1727203858.54172: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203858.58486: done with get_vars() 13731 1727203858.58518: done getting variables 13731 1727203858.58790: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:96 Tuesday 24 September 2024 14:50:58 -0400 (0:00:00.123) 0:00:36.819 ***** 13731 1727203858.58829: entering _queue_task() for managed-node3/package 13731 1727203858.59615: worker is 1 (out of 1 available) 13731 1727203858.59628: exiting _queue_task() for managed-node3/package 13731 1727203858.59640: done queuing things up, now waiting for results queue to drain 13731 1727203858.59641: waiting for pending results... 13731 1727203858.60701: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13731 1727203858.60986: in run() - task 028d2410-947f-82dc-c122-00000000069d 13731 1727203858.60995: variable 'ansible_search_path' from source: unknown 13731 1727203858.61004: variable 'ansible_search_path' from source: unknown 13731 1727203858.61048: calling self._execute() 13731 1727203858.61182: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203858.61298: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203858.61312: variable 'omit' from source: magic vars 13731 1727203858.62254: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.62287: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203858.62422: variable 'network_state' from source: role '' defaults 13731 1727203858.62442: Evaluated conditional (network_state != {}): False 13731 1727203858.62449: when evaluation is False, skipping this task 13731 1727203858.62455: _execute() done 13731 1727203858.62465: dumping result to json 13731 1727203858.62472: done dumping result, returning 13731 1727203858.62496: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable [028d2410-947f-82dc-c122-00000000069d] 13731 1727203858.62507: sending task result for task 028d2410-947f-82dc-c122-00000000069d 13731 1727203858.62751: done sending task result for task 028d2410-947f-82dc-c122-00000000069d 13731 1727203858.62755: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203858.63013: no more pending results, returning what we have 13731 1727203858.63016: results queue empty 13731 1727203858.63017: checking for any_errors_fatal 13731 1727203858.63024: done checking for any_errors_fatal 13731 1727203858.63025: checking for max_fail_percentage 13731 1727203858.63027: done checking for max_fail_percentage 13731 1727203858.63028: checking to see if all hosts have failed and the running result is not ok 13731 1727203858.63028: done checking to see if all hosts have failed 13731 1727203858.63029: getting the remaining hosts for this loop 13731 1727203858.63031: done getting the remaining hosts for this loop 13731 1727203858.63034: getting the next task for host managed-node3 13731 1727203858.63040: done getting next task for host managed-node3 13731 1727203858.63045: ^ task is: TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13731 1727203858.63051: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203858.63068: getting variables 13731 1727203858.63070: in VariableManager get_vars() 13731 1727203858.63110: Calling all_inventory to load vars for managed-node3 13731 1727203858.63113: Calling groups_inventory to load vars for managed-node3 13731 1727203858.63115: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203858.63124: Calling all_plugins_play to load vars for managed-node3 13731 1727203858.63127: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203858.63130: Calling groups_plugins_play to load vars for managed-node3 13731 1727203858.66948: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203858.70167: done with get_vars() 13731 1727203858.70204: done getting variables 13731 1727203858.70267: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:109 Tuesday 24 September 2024 14:50:58 -0400 (0:00:00.114) 0:00:36.934 ***** 13731 1727203858.70308: entering _queue_task() for managed-node3/service 13731 1727203858.70755: worker is 1 (out of 1 available) 13731 1727203858.70768: exiting _queue_task() for managed-node3/service 13731 1727203858.70782: done queuing things up, now waiting for results queue to drain 13731 1727203858.70783: waiting for pending results... 13731 1727203858.71079: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13731 1727203858.71218: in run() - task 028d2410-947f-82dc-c122-00000000069e 13731 1727203858.71237: variable 'ansible_search_path' from source: unknown 13731 1727203858.71241: variable 'ansible_search_path' from source: unknown 13731 1727203858.71279: calling self._execute() 13731 1727203858.71383: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203858.71388: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203858.71390: variable 'omit' from source: magic vars 13731 1727203858.71784: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.71788: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203858.71924: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203858.72116: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203858.75984: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203858.76088: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203858.76132: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203858.76264: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203858.76268: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203858.76332: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.76388: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.76420: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.76467: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.76504: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.76555: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.76606: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.76698: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.76702: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.76707: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.76756: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.76789: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.76911: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.76915: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.76917: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.77151: variable 'network_connections' from source: task vars 13731 1727203858.77173: variable 'port2_profile' from source: play vars 13731 1727203858.77343: variable 'port2_profile' from source: play vars 13731 1727203858.77346: variable 'port1_profile' from source: play vars 13731 1727203858.77347: variable 'port1_profile' from source: play vars 13731 1727203858.77350: variable 'controller_profile' from source: play vars 13731 1727203858.77412: variable 'controller_profile' from source: play vars 13731 1727203858.77564: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203858.77756: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203858.77794: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203858.77884: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203858.77889: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203858.78022: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203858.78062: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203858.78088: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.78117: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203858.78168: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203858.78584: variable 'network_connections' from source: task vars 13731 1727203858.78587: variable 'port2_profile' from source: play vars 13731 1727203858.78645: variable 'port2_profile' from source: play vars 13731 1727203858.78655: variable 'port1_profile' from source: play vars 13731 1727203858.78739: variable 'port1_profile' from source: play vars 13731 1727203858.78782: variable 'controller_profile' from source: play vars 13731 1727203858.78809: variable 'controller_profile' from source: play vars 13731 1727203858.78836: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203858.78847: when evaluation is False, skipping this task 13731 1727203858.78849: _execute() done 13731 1727203858.78852: dumping result to json 13731 1727203858.78854: done dumping result, returning 13731 1727203858.78857: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces [028d2410-947f-82dc-c122-00000000069e] 13731 1727203858.78862: sending task result for task 028d2410-947f-82dc-c122-00000000069e 13731 1727203858.79070: done sending task result for task 028d2410-947f-82dc-c122-00000000069e 13731 1727203858.79074: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203858.79126: no more pending results, returning what we have 13731 1727203858.79131: results queue empty 13731 1727203858.79131: checking for any_errors_fatal 13731 1727203858.79139: done checking for any_errors_fatal 13731 1727203858.79140: checking for max_fail_percentage 13731 1727203858.79142: done checking for max_fail_percentage 13731 1727203858.79143: checking to see if all hosts have failed and the running result is not ok 13731 1727203858.79143: done checking to see if all hosts have failed 13731 1727203858.79144: getting the remaining hosts for this loop 13731 1727203858.79146: done getting the remaining hosts for this loop 13731 1727203858.79149: getting the next task for host managed-node3 13731 1727203858.79157: done getting next task for host managed-node3 13731 1727203858.79162: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13731 1727203858.79167: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203858.79187: getting variables 13731 1727203858.79189: in VariableManager get_vars() 13731 1727203858.79228: Calling all_inventory to load vars for managed-node3 13731 1727203858.79230: Calling groups_inventory to load vars for managed-node3 13731 1727203858.79233: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203858.79243: Calling all_plugins_play to load vars for managed-node3 13731 1727203858.79246: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203858.79249: Calling groups_plugins_play to load vars for managed-node3 13731 1727203858.81051: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203858.84914: done with get_vars() 13731 1727203858.84940: done getting variables 13731 1727203858.85045: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start NetworkManager] ***** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 Tuesday 24 September 2024 14:50:58 -0400 (0:00:00.147) 0:00:37.081 ***** 13731 1727203858.85105: entering _queue_task() for managed-node3/service 13731 1727203858.85705: worker is 1 (out of 1 available) 13731 1727203858.85724: exiting _queue_task() for managed-node3/service 13731 1727203858.85738: done queuing things up, now waiting for results queue to drain 13731 1727203858.85740: waiting for pending results... 13731 1727203858.86182: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13731 1727203858.86300: in run() - task 028d2410-947f-82dc-c122-00000000069f 13731 1727203858.86304: variable 'ansible_search_path' from source: unknown 13731 1727203858.86308: variable 'ansible_search_path' from source: unknown 13731 1727203858.86359: calling self._execute() 13731 1727203858.86437: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203858.86441: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203858.86459: variable 'omit' from source: magic vars 13731 1727203858.86749: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.86786: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203858.86878: variable 'network_provider' from source: set_fact 13731 1727203858.86882: variable 'network_state' from source: role '' defaults 13731 1727203858.86895: Evaluated conditional (network_provider == "nm" or network_state != {}): True 13731 1727203858.86899: variable 'omit' from source: magic vars 13731 1727203858.86948: variable 'omit' from source: magic vars 13731 1727203858.87002: variable 'network_service_name' from source: role '' defaults 13731 1727203858.87043: variable 'network_service_name' from source: role '' defaults 13731 1727203858.87117: variable '__network_provider_setup' from source: role '' defaults 13731 1727203858.87123: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203858.87202: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203858.87206: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203858.87257: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203858.87432: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203858.90644: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203858.90762: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203858.90813: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203858.90880: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203858.91000: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203858.91115: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.91163: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.91208: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.91292: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.91369: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.91420: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.91451: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.91507: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.91615: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.91619: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.92016: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13731 1727203858.92237: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.92241: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.92244: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.92293: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.92485: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.92607: variable 'ansible_python' from source: facts 13731 1727203858.92622: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13731 1727203858.92793: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203858.92878: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203858.93148: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.93151: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.93153: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.93164: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.93186: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.93239: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203858.93298: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203858.93368: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.93380: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203858.93402: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203858.93620: variable 'network_connections' from source: task vars 13731 1727203858.93728: variable 'port2_profile' from source: play vars 13731 1727203858.93750: variable 'port2_profile' from source: play vars 13731 1727203858.93773: variable 'port1_profile' from source: play vars 13731 1727203858.93877: variable 'port1_profile' from source: play vars 13731 1727203858.93957: variable 'controller_profile' from source: play vars 13731 1727203858.94187: variable 'controller_profile' from source: play vars 13731 1727203858.94509: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203858.94634: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203858.94734: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203858.94790: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203858.94848: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203858.94954: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203858.94995: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203858.95037: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203858.95122: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203858.95204: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203858.95836: variable 'network_connections' from source: task vars 13731 1727203858.95855: variable 'port2_profile' from source: play vars 13731 1727203858.96362: variable 'port2_profile' from source: play vars 13731 1727203858.96366: variable 'port1_profile' from source: play vars 13731 1727203858.96369: variable 'port1_profile' from source: play vars 13731 1727203858.96371: variable 'controller_profile' from source: play vars 13731 1727203858.96407: variable 'controller_profile' from source: play vars 13731 1727203858.96459: variable '__network_packages_default_wireless' from source: role '' defaults 13731 1727203858.96635: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203858.97287: variable 'network_connections' from source: task vars 13731 1727203858.97320: variable 'port2_profile' from source: play vars 13731 1727203858.97457: variable 'port2_profile' from source: play vars 13731 1727203858.97478: variable 'port1_profile' from source: play vars 13731 1727203858.97633: variable 'port1_profile' from source: play vars 13731 1727203858.97683: variable 'controller_profile' from source: play vars 13731 1727203858.97788: variable 'controller_profile' from source: play vars 13731 1727203858.97828: variable '__network_packages_default_team' from source: role '' defaults 13731 1727203858.97930: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203858.98282: variable 'network_connections' from source: task vars 13731 1727203858.98293: variable 'port2_profile' from source: play vars 13731 1727203858.98320: variable 'port2_profile' from source: play vars 13731 1727203858.98327: variable 'port1_profile' from source: play vars 13731 1727203858.98411: variable 'port1_profile' from source: play vars 13731 1727203858.98414: variable 'controller_profile' from source: play vars 13731 1727203858.98481: variable 'controller_profile' from source: play vars 13731 1727203858.98554: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203858.98593: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203858.98596: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203858.98674: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203858.98900: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13731 1727203858.99527: variable 'network_connections' from source: task vars 13731 1727203858.99530: variable 'port2_profile' from source: play vars 13731 1727203858.99533: variable 'port2_profile' from source: play vars 13731 1727203858.99536: variable 'port1_profile' from source: play vars 13731 1727203858.99590: variable 'port1_profile' from source: play vars 13731 1727203858.99598: variable 'controller_profile' from source: play vars 13731 1727203858.99662: variable 'controller_profile' from source: play vars 13731 1727203858.99665: variable 'ansible_distribution' from source: facts 13731 1727203858.99673: variable '__network_rh_distros' from source: role '' defaults 13731 1727203858.99688: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.99728: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13731 1727203858.99889: variable 'ansible_distribution' from source: facts 13731 1727203858.99893: variable '__network_rh_distros' from source: role '' defaults 13731 1727203858.99954: variable 'ansible_distribution_major_version' from source: facts 13731 1727203858.99958: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13731 1727203859.00148: variable 'ansible_distribution' from source: facts 13731 1727203859.00184: variable '__network_rh_distros' from source: role '' defaults 13731 1727203859.00187: variable 'ansible_distribution_major_version' from source: facts 13731 1727203859.00189: variable 'network_provider' from source: set_fact 13731 1727203859.00225: variable 'omit' from source: magic vars 13731 1727203859.00228: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203859.00302: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203859.00305: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203859.00307: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203859.00315: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203859.00383: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203859.00387: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203859.00390: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203859.00480: Set connection var ansible_pipelining to False 13731 1727203859.00483: Set connection var ansible_shell_type to sh 13731 1727203859.00489: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203859.00492: Set connection var ansible_connection to ssh 13731 1727203859.00494: Set connection var ansible_shell_executable to /bin/sh 13731 1727203859.00496: Set connection var ansible_timeout to 10 13731 1727203859.00530: variable 'ansible_shell_executable' from source: unknown 13731 1727203859.00536: variable 'ansible_connection' from source: unknown 13731 1727203859.00538: variable 'ansible_module_compression' from source: unknown 13731 1727203859.00540: variable 'ansible_shell_type' from source: unknown 13731 1727203859.00542: variable 'ansible_shell_executable' from source: unknown 13731 1727203859.00544: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203859.00546: variable 'ansible_pipelining' from source: unknown 13731 1727203859.00548: variable 'ansible_timeout' from source: unknown 13731 1727203859.00550: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203859.00705: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203859.00709: variable 'omit' from source: magic vars 13731 1727203859.00712: starting attempt loop 13731 1727203859.00714: running the handler 13731 1727203859.00736: variable 'ansible_facts' from source: unknown 13731 1727203859.01504: _low_level_execute_command(): starting 13731 1727203859.01507: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203859.02196: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203859.02248: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203859.02281: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203859.02372: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203859.02519: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203859.02526: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203859.04158: stdout chunk (state=3): >>>/root <<< 13731 1727203859.04346: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203859.04355: stdout chunk (state=3): >>><<< 13731 1727203859.04358: stderr chunk (state=3): >>><<< 13731 1727203859.04416: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203859.04420: _low_level_execute_command(): starting 13731 1727203859.04521: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598 `" && echo ansible-tmp-1727203859.0439386-16286-5444021159598="` echo /root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598 `" ) && sleep 0' 13731 1727203859.05021: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203859.05027: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203859.05053: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.05057: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203859.05062: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.05115: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203859.05153: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203859.05201: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203859.07116: stdout chunk (state=3): >>>ansible-tmp-1727203859.0439386-16286-5444021159598=/root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598 <<< 13731 1727203859.07234: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203859.07284: stderr chunk (state=3): >>><<< 13731 1727203859.07288: stdout chunk (state=3): >>><<< 13731 1727203859.07291: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203859.0439386-16286-5444021159598=/root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203859.07516: variable 'ansible_module_compression' from source: unknown 13731 1727203859.07519: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.systemd-ZIP_DEFLATED 13731 1727203859.07951: variable 'ansible_facts' from source: unknown 13731 1727203859.08449: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/AnsiballZ_systemd.py 13731 1727203859.08901: Sending initial data 13731 1727203859.08912: Sent initial data (154 bytes) 13731 1727203859.09384: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203859.09398: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.09409: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.09456: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203859.09483: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203859.09510: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203859.11009: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 <<< 13731 1727203859.11012: stderr chunk (state=3): >>>debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203859.11038: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203859.11073: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp05levmd9 /root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/AnsiballZ_systemd.py <<< 13731 1727203859.11078: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/AnsiballZ_systemd.py" <<< 13731 1727203859.11109: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp05levmd9" to remote "/root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/AnsiballZ_systemd.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/AnsiballZ_systemd.py" <<< 13731 1727203859.12909: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203859.12912: stderr chunk (state=3): >>><<< 13731 1727203859.12915: stdout chunk (state=3): >>><<< 13731 1727203859.12931: done transferring module to remote 13731 1727203859.12941: _low_level_execute_command(): starting 13731 1727203859.12944: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/ /root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/AnsiballZ_systemd.py && sleep 0' 13731 1727203859.14166: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203859.14170: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203859.14173: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203859.14189: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203859.14201: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203859.14209: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203859.14218: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.14232: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203859.14602: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203859.14623: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203859.16379: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203859.16383: stdout chunk (state=3): >>><<< 13731 1727203859.16390: stderr chunk (state=3): >>><<< 13731 1727203859.16406: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203859.16409: _low_level_execute_command(): starting 13731 1727203859.16413: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/AnsiballZ_systemd.py && sleep 0' 13731 1727203859.17331: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203859.17364: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203859.17383: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203859.17486: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.17521: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203859.17543: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203859.17569: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203859.17611: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203859.46345: stdout chunk (state=3): >>> {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "RestartMode": "normal", "NotifyAccess": "none", "RestartUSec": "100ms", "RestartSteps": "0", "RestartMaxDelayUSec": "infinity", "RestartUSecNext": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "704", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "FileDescriptorStorePreserve": "restart", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainStartTimestampMonotonic": "28990148", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainHandoffTimestampMonotonic": "29005881", "ExecMainPID": "704", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "2938", "MemoryCurrent": "10358784", "MemoryPeak": "13586432", "MemorySwapCurrent": "0", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryAvailable": "3284160512", "EffectiveMemoryMax": "3702865920", "EffectiveMemoryHigh": "3702865920", "CPUUsageNSec": "660867000", "TasksCurrent": "4", "EffectiveTasksMax": "22362", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultStartupMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "StartupMemoryLow": "0", "MemoryHigh": "infinity", "StartupMemoryHigh": "infinity", "MemoryMax": "infinity", "StartupMemoryMax": "infinity", "MemorySwapMax": "infinity", "StartupMemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22362", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "MemoryPressureWatch": "auto", "MemoryPressureThresholdUSec": "200ms", "CoredumpRe<<< 13731 1727203859.46392: stdout chunk (state=3): >>>ceive": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "RootEphemeral": "no", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "SetLoginEnvironment": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "MemoryKSM": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "system.slice dbus.socket sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "cloud-init.service shutdown.target NetworkManager-wait-online.service multi-user.target network.target", "After": "dbus-br<<< 13731 1727203859.46401: stdout chunk (state=3): >>>oker.service systemd-journald.socket network-pre.target basic.target cloud-init-local.service dbus.socket system.slice sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:49:45 EDT", "StateChangeTimestampMonotonic": "362725592", "InactiveExitTimestamp": "Tue 2024-09-24 14:44:11 EDT", "InactiveExitTimestampMonotonic": "28990654", "ActiveEnterTimestamp": "Tue 2024-09-24 14:44:12 EDT", "ActiveEnterTimestampMonotonic": "29769382", "ActiveExitTimestampMonotonic": "0", "InactiveEnterTimestampMonotonic": "0", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "SurviveFinalKillSignal": "no", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ConditionTimestampMonotonic": "28989295", "AssertTimestamp": "Tue 2024-09-24 14:44:11 EDT", "AssertTimestampMonotonic": "28989297", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "70a845f8a1964db89963090ed497f47f", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 13731 1727203859.48167: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203859.48218: stderr chunk (state=3): >>><<< 13731 1727203859.48222: stdout chunk (state=3): >>><<< 13731 1727203859.48256: _low_level_execute_command() done: rc=0, stdout= {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "RestartMode": "normal", "NotifyAccess": "none", "RestartUSec": "100ms", "RestartSteps": "0", "RestartMaxDelayUSec": "infinity", "RestartUSecNext": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "704", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "FileDescriptorStorePreserve": "restart", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainStartTimestampMonotonic": "28990148", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainHandoffTimestampMonotonic": "29005881", "ExecMainPID": "704", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "2938", "MemoryCurrent": "10358784", "MemoryPeak": "13586432", "MemorySwapCurrent": "0", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryAvailable": "3284160512", "EffectiveMemoryMax": "3702865920", "EffectiveMemoryHigh": "3702865920", "CPUUsageNSec": "660867000", "TasksCurrent": "4", "EffectiveTasksMax": "22362", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultStartupMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "StartupMemoryLow": "0", "MemoryHigh": "infinity", "StartupMemoryHigh": "infinity", "MemoryMax": "infinity", "StartupMemoryMax": "infinity", "MemorySwapMax": "infinity", "StartupMemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22362", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "MemoryPressureWatch": "auto", "MemoryPressureThresholdUSec": "200ms", "CoredumpReceive": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "RootEphemeral": "no", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "SetLoginEnvironment": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "MemoryKSM": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "system.slice dbus.socket sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "cloud-init.service shutdown.target NetworkManager-wait-online.service multi-user.target network.target", "After": "dbus-broker.service systemd-journald.socket network-pre.target basic.target cloud-init-local.service dbus.socket system.slice sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:49:45 EDT", "StateChangeTimestampMonotonic": "362725592", "InactiveExitTimestamp": "Tue 2024-09-24 14:44:11 EDT", "InactiveExitTimestampMonotonic": "28990654", "ActiveEnterTimestamp": "Tue 2024-09-24 14:44:12 EDT", "ActiveEnterTimestampMonotonic": "29769382", "ActiveExitTimestampMonotonic": "0", "InactiveEnterTimestampMonotonic": "0", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "SurviveFinalKillSignal": "no", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ConditionTimestampMonotonic": "28989295", "AssertTimestamp": "Tue 2024-09-24 14:44:11 EDT", "AssertTimestampMonotonic": "28989297", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "70a845f8a1964db89963090ed497f47f", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203859.48385: done with _execute_module (ansible.legacy.systemd, {'name': 'NetworkManager', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203859.48400: _low_level_execute_command(): starting 13731 1727203859.48404: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203859.0439386-16286-5444021159598/ > /dev/null 2>&1 && sleep 0' 13731 1727203859.48916: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203859.48920: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.48922: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203859.48924: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.48970: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203859.48984: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203859.49019: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203859.50787: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203859.50833: stderr chunk (state=3): >>><<< 13731 1727203859.50836: stdout chunk (state=3): >>><<< 13731 1727203859.50883: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203859.50886: handler run complete 13731 1727203859.50944: attempt loop complete, returning result 13731 1727203859.50949: _execute() done 13731 1727203859.50952: dumping result to json 13731 1727203859.50983: done dumping result, returning 13731 1727203859.50986: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager [028d2410-947f-82dc-c122-00000000069f] 13731 1727203859.50988: sending task result for task 028d2410-947f-82dc-c122-00000000069f ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203859.51426: no more pending results, returning what we have 13731 1727203859.51429: results queue empty 13731 1727203859.51430: checking for any_errors_fatal 13731 1727203859.51434: done checking for any_errors_fatal 13731 1727203859.51435: checking for max_fail_percentage 13731 1727203859.51437: done checking for max_fail_percentage 13731 1727203859.51437: checking to see if all hosts have failed and the running result is not ok 13731 1727203859.51438: done checking to see if all hosts have failed 13731 1727203859.51444: getting the remaining hosts for this loop 13731 1727203859.51446: done getting the remaining hosts for this loop 13731 1727203859.51449: getting the next task for host managed-node3 13731 1727203859.51488: done getting next task for host managed-node3 13731 1727203859.51492: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13731 1727203859.51497: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203859.51505: done sending task result for task 028d2410-947f-82dc-c122-00000000069f 13731 1727203859.51507: WORKER PROCESS EXITING 13731 1727203859.51515: getting variables 13731 1727203859.51516: in VariableManager get_vars() 13731 1727203859.51545: Calling all_inventory to load vars for managed-node3 13731 1727203859.51552: Calling groups_inventory to load vars for managed-node3 13731 1727203859.51554: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203859.51565: Calling all_plugins_play to load vars for managed-node3 13731 1727203859.51567: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203859.51569: Calling groups_plugins_play to load vars for managed-node3 13731 1727203859.52601: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203859.53629: done with get_vars() 13731 1727203859.53646: done getting variables 13731 1727203859.53695: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start wpa_supplicant] ***** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:133 Tuesday 24 September 2024 14:50:59 -0400 (0:00:00.686) 0:00:37.768 ***** 13731 1727203859.53723: entering _queue_task() for managed-node3/service 13731 1727203859.54009: worker is 1 (out of 1 available) 13731 1727203859.54026: exiting _queue_task() for managed-node3/service 13731 1727203859.54038: done queuing things up, now waiting for results queue to drain 13731 1727203859.54040: waiting for pending results... 13731 1727203859.54241: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13731 1727203859.54402: in run() - task 028d2410-947f-82dc-c122-0000000006a0 13731 1727203859.54443: variable 'ansible_search_path' from source: unknown 13731 1727203859.54446: variable 'ansible_search_path' from source: unknown 13731 1727203859.54464: calling self._execute() 13731 1727203859.54535: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203859.54539: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203859.54550: variable 'omit' from source: magic vars 13731 1727203859.54851: variable 'ansible_distribution_major_version' from source: facts 13731 1727203859.54870: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203859.54950: variable 'network_provider' from source: set_fact 13731 1727203859.54954: Evaluated conditional (network_provider == "nm"): True 13731 1727203859.55022: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203859.55101: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203859.55219: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203859.57125: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203859.57149: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203859.57189: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203859.57228: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203859.57245: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203859.57322: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203859.57354: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203859.57377: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203859.57416: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203859.57429: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203859.57478: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203859.57499: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203859.57522: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203859.57559: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203859.57583: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203859.57664: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203859.57667: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203859.57678: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203859.57698: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203859.57713: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203859.57855: variable 'network_connections' from source: task vars 13731 1727203859.57882: variable 'port2_profile' from source: play vars 13731 1727203859.57935: variable 'port2_profile' from source: play vars 13731 1727203859.57943: variable 'port1_profile' from source: play vars 13731 1727203859.57999: variable 'port1_profile' from source: play vars 13731 1727203859.58003: variable 'controller_profile' from source: play vars 13731 1727203859.58054: variable 'controller_profile' from source: play vars 13731 1727203859.58124: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203859.58245: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203859.58274: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203859.58298: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203859.58321: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203859.58380: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203859.58406: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203859.58439: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203859.58470: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203859.58527: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203859.58789: variable 'network_connections' from source: task vars 13731 1727203859.58799: variable 'port2_profile' from source: play vars 13731 1727203859.58855: variable 'port2_profile' from source: play vars 13731 1727203859.58861: variable 'port1_profile' from source: play vars 13731 1727203859.58926: variable 'port1_profile' from source: play vars 13731 1727203859.58937: variable 'controller_profile' from source: play vars 13731 1727203859.59029: variable 'controller_profile' from source: play vars 13731 1727203859.59032: Evaluated conditional (__network_wpa_supplicant_required): False 13731 1727203859.59035: when evaluation is False, skipping this task 13731 1727203859.59053: _execute() done 13731 1727203859.59056: dumping result to json 13731 1727203859.59058: done dumping result, returning 13731 1727203859.59063: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant [028d2410-947f-82dc-c122-0000000006a0] 13731 1727203859.59065: sending task result for task 028d2410-947f-82dc-c122-0000000006a0 13731 1727203859.59178: done sending task result for task 028d2410-947f-82dc-c122-0000000006a0 13731 1727203859.59181: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wpa_supplicant_required", "skip_reason": "Conditional result was False" } 13731 1727203859.59243: no more pending results, returning what we have 13731 1727203859.59248: results queue empty 13731 1727203859.59249: checking for any_errors_fatal 13731 1727203859.59269: done checking for any_errors_fatal 13731 1727203859.59270: checking for max_fail_percentage 13731 1727203859.59272: done checking for max_fail_percentage 13731 1727203859.59273: checking to see if all hosts have failed and the running result is not ok 13731 1727203859.59274: done checking to see if all hosts have failed 13731 1727203859.59274: getting the remaining hosts for this loop 13731 1727203859.59278: done getting the remaining hosts for this loop 13731 1727203859.59290: getting the next task for host managed-node3 13731 1727203859.59299: done getting next task for host managed-node3 13731 1727203859.59303: ^ task is: TASK: fedora.linux_system_roles.network : Enable network service 13731 1727203859.59311: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203859.59328: getting variables 13731 1727203859.59330: in VariableManager get_vars() 13731 1727203859.59365: Calling all_inventory to load vars for managed-node3 13731 1727203859.59368: Calling groups_inventory to load vars for managed-node3 13731 1727203859.59373: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203859.59386: Calling all_plugins_play to load vars for managed-node3 13731 1727203859.59389: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203859.59392: Calling groups_plugins_play to load vars for managed-node3 13731 1727203859.60738: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203859.62373: done with get_vars() 13731 1727203859.62392: done getting variables 13731 1727203859.62435: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable network service] ************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:142 Tuesday 24 September 2024 14:50:59 -0400 (0:00:00.087) 0:00:37.855 ***** 13731 1727203859.62463: entering _queue_task() for managed-node3/service 13731 1727203859.62717: worker is 1 (out of 1 available) 13731 1727203859.62732: exiting _queue_task() for managed-node3/service 13731 1727203859.62745: done queuing things up, now waiting for results queue to drain 13731 1727203859.62747: waiting for pending results... 13731 1727203859.62930: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable network service 13731 1727203859.63040: in run() - task 028d2410-947f-82dc-c122-0000000006a1 13731 1727203859.63050: variable 'ansible_search_path' from source: unknown 13731 1727203859.63054: variable 'ansible_search_path' from source: unknown 13731 1727203859.63094: calling self._execute() 13731 1727203859.63172: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203859.63177: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203859.63187: variable 'omit' from source: magic vars 13731 1727203859.63468: variable 'ansible_distribution_major_version' from source: facts 13731 1727203859.63478: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203859.63557: variable 'network_provider' from source: set_fact 13731 1727203859.63560: Evaluated conditional (network_provider == "initscripts"): False 13731 1727203859.63567: when evaluation is False, skipping this task 13731 1727203859.63570: _execute() done 13731 1727203859.63572: dumping result to json 13731 1727203859.63575: done dumping result, returning 13731 1727203859.63584: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable network service [028d2410-947f-82dc-c122-0000000006a1] 13731 1727203859.63589: sending task result for task 028d2410-947f-82dc-c122-0000000006a1 13731 1727203859.63673: done sending task result for task 028d2410-947f-82dc-c122-0000000006a1 13731 1727203859.63677: WORKER PROCESS EXITING skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203859.63720: no more pending results, returning what we have 13731 1727203859.63724: results queue empty 13731 1727203859.63725: checking for any_errors_fatal 13731 1727203859.63735: done checking for any_errors_fatal 13731 1727203859.63736: checking for max_fail_percentage 13731 1727203859.63737: done checking for max_fail_percentage 13731 1727203859.63739: checking to see if all hosts have failed and the running result is not ok 13731 1727203859.63739: done checking to see if all hosts have failed 13731 1727203859.63740: getting the remaining hosts for this loop 13731 1727203859.63742: done getting the remaining hosts for this loop 13731 1727203859.63745: getting the next task for host managed-node3 13731 1727203859.63753: done getting next task for host managed-node3 13731 1727203859.63757: ^ task is: TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13731 1727203859.63762: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=19, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203859.63783: getting variables 13731 1727203859.63784: in VariableManager get_vars() 13731 1727203859.63816: Calling all_inventory to load vars for managed-node3 13731 1727203859.63819: Calling groups_inventory to load vars for managed-node3 13731 1727203859.63820: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203859.63829: Calling all_plugins_play to load vars for managed-node3 13731 1727203859.63832: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203859.63834: Calling groups_plugins_play to load vars for managed-node3 13731 1727203859.64727: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203859.66190: done with get_vars() 13731 1727203859.66215: done getting variables 13731 1727203859.66279: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ensure initscripts network file dependency is present] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:150 Tuesday 24 September 2024 14:50:59 -0400 (0:00:00.038) 0:00:37.894 ***** 13731 1727203859.66316: entering _queue_task() for managed-node3/copy 13731 1727203859.66663: worker is 1 (out of 1 available) 13731 1727203859.66682: exiting _queue_task() for managed-node3/copy 13731 1727203859.66695: done queuing things up, now waiting for results queue to drain 13731 1727203859.66697: waiting for pending results... 13731 1727203859.67099: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13731 1727203859.67181: in run() - task 028d2410-947f-82dc-c122-0000000006a2 13731 1727203859.67231: variable 'ansible_search_path' from source: unknown 13731 1727203859.67240: variable 'ansible_search_path' from source: unknown 13731 1727203859.67243: calling self._execute() 13731 1727203859.67336: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203859.67340: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203859.67384: variable 'omit' from source: magic vars 13731 1727203859.67712: variable 'ansible_distribution_major_version' from source: facts 13731 1727203859.67723: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203859.67834: variable 'network_provider' from source: set_fact 13731 1727203859.67844: Evaluated conditional (network_provider == "initscripts"): False 13731 1727203859.67847: when evaluation is False, skipping this task 13731 1727203859.67849: _execute() done 13731 1727203859.67933: dumping result to json 13731 1727203859.67936: done dumping result, returning 13731 1727203859.67939: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present [028d2410-947f-82dc-c122-0000000006a2] 13731 1727203859.67941: sending task result for task 028d2410-947f-82dc-c122-0000000006a2 13731 1727203859.68140: done sending task result for task 028d2410-947f-82dc-c122-0000000006a2 13731 1727203859.68143: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_provider == \"initscripts\"", "skip_reason": "Conditional result was False" } 13731 1727203859.68184: no more pending results, returning what we have 13731 1727203859.68187: results queue empty 13731 1727203859.68188: checking for any_errors_fatal 13731 1727203859.68194: done checking for any_errors_fatal 13731 1727203859.68195: checking for max_fail_percentage 13731 1727203859.68196: done checking for max_fail_percentage 13731 1727203859.68197: checking to see if all hosts have failed and the running result is not ok 13731 1727203859.68198: done checking to see if all hosts have failed 13731 1727203859.68198: getting the remaining hosts for this loop 13731 1727203859.68200: done getting the remaining hosts for this loop 13731 1727203859.68203: getting the next task for host managed-node3 13731 1727203859.68214: done getting next task for host managed-node3 13731 1727203859.68218: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13731 1727203859.68228: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=20, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203859.68248: getting variables 13731 1727203859.68250: in VariableManager get_vars() 13731 1727203859.68289: Calling all_inventory to load vars for managed-node3 13731 1727203859.68292: Calling groups_inventory to load vars for managed-node3 13731 1727203859.68297: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203859.68306: Calling all_plugins_play to load vars for managed-node3 13731 1727203859.68309: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203859.68312: Calling groups_plugins_play to load vars for managed-node3 13731 1727203859.69829: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203859.71424: done with get_vars() 13731 1727203859.71451: done getting variables TASK [fedora.linux_system_roles.network : Configure networking connection profiles] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 Tuesday 24 September 2024 14:50:59 -0400 (0:00:00.052) 0:00:37.946 ***** 13731 1727203859.71557: entering _queue_task() for managed-node3/fedora.linux_system_roles.network_connections 13731 1727203859.71980: worker is 1 (out of 1 available) 13731 1727203859.71993: exiting _queue_task() for managed-node3/fedora.linux_system_roles.network_connections 13731 1727203859.72011: done queuing things up, now waiting for results queue to drain 13731 1727203859.72012: waiting for pending results... 13731 1727203859.72390: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13731 1727203859.72562: in run() - task 028d2410-947f-82dc-c122-0000000006a3 13731 1727203859.72579: variable 'ansible_search_path' from source: unknown 13731 1727203859.72583: variable 'ansible_search_path' from source: unknown 13731 1727203859.72634: calling self._execute() 13731 1727203859.72774: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203859.72779: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203859.72782: variable 'omit' from source: magic vars 13731 1727203859.73258: variable 'ansible_distribution_major_version' from source: facts 13731 1727203859.73261: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203859.73267: variable 'omit' from source: magic vars 13731 1727203859.73348: variable 'omit' from source: magic vars 13731 1727203859.73576: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203859.76285: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203859.76406: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203859.76500: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203859.76504: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203859.76542: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203859.76680: variable 'network_provider' from source: set_fact 13731 1727203859.76826: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203859.76881: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203859.76905: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203859.76943: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203859.76957: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203859.77043: variable 'omit' from source: magic vars 13731 1727203859.77181: variable 'omit' from source: magic vars 13731 1727203859.77256: variable 'network_connections' from source: task vars 13731 1727203859.77299: variable 'port2_profile' from source: play vars 13731 1727203859.77330: variable 'port2_profile' from source: play vars 13731 1727203859.77374: variable 'port1_profile' from source: play vars 13731 1727203859.77400: variable 'port1_profile' from source: play vars 13731 1727203859.77413: variable 'controller_profile' from source: play vars 13731 1727203859.77581: variable 'controller_profile' from source: play vars 13731 1727203859.77632: variable 'omit' from source: magic vars 13731 1727203859.77640: variable '__lsr_ansible_managed' from source: task vars 13731 1727203859.77701: variable '__lsr_ansible_managed' from source: task vars 13731 1727203859.77901: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 13731 1727203859.78136: Loaded config def from plugin (lookup/template) 13731 1727203859.78139: Loading LookupModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/template.py 13731 1727203859.78173: File lookup term: get_ansible_managed.j2 13731 1727203859.78179: variable 'ansible_search_path' from source: unknown 13731 1727203859.78184: evaluation_path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks 13731 1727203859.78199: search_path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 13731 1727203859.78215: variable 'ansible_search_path' from source: unknown 13731 1727203859.90670: variable 'ansible_managed' from source: unknown 13731 1727203859.91084: variable 'omit' from source: magic vars 13731 1727203859.91088: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203859.91194: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203859.91286: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203859.91317: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203859.91481: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203859.91485: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203859.91487: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203859.91489: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203859.91699: Set connection var ansible_pipelining to False 13731 1727203859.91745: Set connection var ansible_shell_type to sh 13731 1727203859.91756: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203859.91824: Set connection var ansible_connection to ssh 13731 1727203859.91854: Set connection var ansible_shell_executable to /bin/sh 13731 1727203859.91934: Set connection var ansible_timeout to 10 13731 1727203859.91979: variable 'ansible_shell_executable' from source: unknown 13731 1727203859.92281: variable 'ansible_connection' from source: unknown 13731 1727203859.92284: variable 'ansible_module_compression' from source: unknown 13731 1727203859.92286: variable 'ansible_shell_type' from source: unknown 13731 1727203859.92288: variable 'ansible_shell_executable' from source: unknown 13731 1727203859.92290: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203859.92291: variable 'ansible_pipelining' from source: unknown 13731 1727203859.92293: variable 'ansible_timeout' from source: unknown 13731 1727203859.92301: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203859.92682: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203859.92686: variable 'omit' from source: magic vars 13731 1727203859.92688: starting attempt loop 13731 1727203859.92691: running the handler 13731 1727203859.92693: _low_level_execute_command(): starting 13731 1727203859.92696: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203859.94092: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203859.94246: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203859.94263: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203859.94521: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203859.94723: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203859.96771: stdout chunk (state=3): >>>/root <<< 13731 1727203859.96778: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203859.96780: stdout chunk (state=3): >>><<< 13731 1727203859.96782: stderr chunk (state=3): >>><<< 13731 1727203859.96785: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203859.96787: _low_level_execute_command(): starting 13731 1727203859.96789: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126 `" && echo ansible-tmp-1727203859.9667153-16337-72342509467126="` echo /root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126 `" ) && sleep 0' 13731 1727203859.98179: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203859.98184: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.98187: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203859.98189: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203859.98191: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203859.98982: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203860.00668: stdout chunk (state=3): >>>ansible-tmp-1727203859.9667153-16337-72342509467126=/root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126 <<< 13731 1727203860.00725: stdout chunk (state=3): >>><<< 13731 1727203860.00734: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203860.00743: stderr chunk (state=3): >>><<< 13731 1727203860.00883: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203859.9667153-16337-72342509467126=/root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203860.01081: variable 'ansible_module_compression' from source: unknown 13731 1727203860.01202: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.network_connections-ZIP_DEFLATED 13731 1727203860.01471: variable 'ansible_facts' from source: unknown 13731 1727203860.01477: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/AnsiballZ_network_connections.py 13731 1727203860.01999: Sending initial data 13731 1727203860.02010: Sent initial data (167 bytes) 13731 1727203860.03146: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203860.03310: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203860.03556: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203860.03996: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203860.05198: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203860.05244: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203860.05302: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmppvk5d53x /root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/AnsiballZ_network_connections.py <<< 13731 1727203860.05312: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/AnsiballZ_network_connections.py" <<< 13731 1727203860.05516: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmppvk5d53x" to remote "/root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/AnsiballZ_network_connections.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/AnsiballZ_network_connections.py" <<< 13731 1727203860.07613: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203860.07617: stdout chunk (state=3): >>><<< 13731 1727203860.07620: stderr chunk (state=3): >>><<< 13731 1727203860.07622: done transferring module to remote 13731 1727203860.07624: _low_level_execute_command(): starting 13731 1727203860.07626: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/ /root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/AnsiballZ_network_connections.py && sleep 0' 13731 1727203860.09084: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203860.09198: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203860.09219: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203860.09293: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203860.11057: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203860.11072: stdout chunk (state=3): >>><<< 13731 1727203860.11086: stderr chunk (state=3): >>><<< 13731 1727203860.11260: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203860.11557: _low_level_execute_command(): starting 13731 1727203860.11561: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/AnsiballZ_network_connections.py && sleep 0' 13731 1727203860.13880: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203860.14007: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203860.14424: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203860.66563: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.1/52cf19c0-ba8d-4c54-a280-415ff537447d: error=unknown <<< 13731 1727203860.68263: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.0/4dc9735e-3039-485e-9eca-4fa95cb0bf9d: error=unknown <<< 13731 1727203860.69996: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0/49770e6d-8624-49fd-b707-6e339710b2f8: error=unknown <<< 13731 1727203860.70228: stdout chunk (state=3): >>> {"changed": true, "warnings": [], "stderr": "\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} <<< 13731 1727203860.72217: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203860.72223: stdout chunk (state=3): >>><<< 13731 1727203860.72226: stderr chunk (state=3): >>><<< 13731 1727203860.72518: _low_level_execute_command() done: rc=0, stdout=Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.1/52cf19c0-ba8d-4c54-a280-415ff537447d: error=unknown Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.0/4dc9735e-3039-485e-9eca-4fa95cb0bf9d: error=unknown Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_4s2gjksi/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0/49770e6d-8624-49fd-b707-6e339710b2f8: error=unknown {"changed": true, "warnings": [], "stderr": "\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203860.72523: done with _execute_module (fedora.linux_system_roles.network_connections, {'provider': 'nm', 'connections': [{'name': 'bond0.1', 'persistent_state': 'absent', 'state': 'down'}, {'name': 'bond0.0', 'persistent_state': 'absent', 'state': 'down'}, {'name': 'bond0', 'persistent_state': 'absent', 'state': 'down'}], '__header': '#\n# Ansible managed\n#\n# system_role:network\n', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'fedora.linux_system_roles.network_connections', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203860.72529: _low_level_execute_command(): starting 13731 1727203860.72532: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203859.9667153-16337-72342509467126/ > /dev/null 2>&1 && sleep 0' 13731 1727203860.73802: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203860.73894: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203860.73912: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203860.73928: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203860.73945: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203860.74071: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203860.74221: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203860.74254: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203860.76130: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203860.76381: stderr chunk (state=3): >>><<< 13731 1727203860.76385: stdout chunk (state=3): >>><<< 13731 1727203860.76388: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203860.76390: handler run complete 13731 1727203860.76393: attempt loop complete, returning result 13731 1727203860.76395: _execute() done 13731 1727203860.76397: dumping result to json 13731 1727203860.76399: done dumping result, returning 13731 1727203860.76401: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking connection profiles [028d2410-947f-82dc-c122-0000000006a3] 13731 1727203860.76403: sending task result for task 028d2410-947f-82dc-c122-0000000006a3 changed: [managed-node3] => { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "name": "bond0.1", "persistent_state": "absent", "state": "down" }, { "name": "bond0.0", "persistent_state": "absent", "state": "down" }, { "name": "bond0", "persistent_state": "absent", "state": "down" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true } STDERR: 13731 1727203860.76916: no more pending results, returning what we have 13731 1727203860.76921: results queue empty 13731 1727203860.76922: checking for any_errors_fatal 13731 1727203860.76931: done checking for any_errors_fatal 13731 1727203860.76932: checking for max_fail_percentage 13731 1727203860.76934: done checking for max_fail_percentage 13731 1727203860.76935: checking to see if all hosts have failed and the running result is not ok 13731 1727203860.76935: done checking to see if all hosts have failed 13731 1727203860.76936: getting the remaining hosts for this loop 13731 1727203860.76939: done getting the remaining hosts for this loop 13731 1727203860.76942: getting the next task for host managed-node3 13731 1727203860.76950: done getting next task for host managed-node3 13731 1727203860.76954: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking state 13731 1727203860.76958: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=21, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203860.76971: getting variables 13731 1727203860.76973: in VariableManager get_vars() 13731 1727203860.77264: Calling all_inventory to load vars for managed-node3 13731 1727203860.77267: Calling groups_inventory to load vars for managed-node3 13731 1727203860.77270: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203860.77282: Calling all_plugins_play to load vars for managed-node3 13731 1727203860.77284: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203860.77288: Calling groups_plugins_play to load vars for managed-node3 13731 1727203860.77808: done sending task result for task 028d2410-947f-82dc-c122-0000000006a3 13731 1727203860.77812: WORKER PROCESS EXITING 13731 1727203860.80417: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203860.83338: done with get_vars() 13731 1727203860.83369: done getting variables TASK [fedora.linux_system_roles.network : Configure networking state] ********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:171 Tuesday 24 September 2024 14:51:00 -0400 (0:00:01.119) 0:00:39.066 ***** 13731 1727203860.83509: entering _queue_task() for managed-node3/fedora.linux_system_roles.network_state 13731 1727203860.84360: worker is 1 (out of 1 available) 13731 1727203860.84373: exiting _queue_task() for managed-node3/fedora.linux_system_roles.network_state 13731 1727203860.84388: done queuing things up, now waiting for results queue to drain 13731 1727203860.84390: waiting for pending results... 13731 1727203860.84922: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking state 13731 1727203860.85086: in run() - task 028d2410-947f-82dc-c122-0000000006a4 13731 1727203860.85102: variable 'ansible_search_path' from source: unknown 13731 1727203860.85105: variable 'ansible_search_path' from source: unknown 13731 1727203860.85142: calling self._execute() 13731 1727203860.85240: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203860.85244: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203860.85253: variable 'omit' from source: magic vars 13731 1727203860.85726: variable 'ansible_distribution_major_version' from source: facts 13731 1727203860.85756: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203860.85909: variable 'network_state' from source: role '' defaults 13731 1727203860.85914: Evaluated conditional (network_state != {}): False 13731 1727203860.85917: when evaluation is False, skipping this task 13731 1727203860.85919: _execute() done 13731 1727203860.85922: dumping result to json 13731 1727203860.85925: done dumping result, returning 13731 1727203860.85928: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking state [028d2410-947f-82dc-c122-0000000006a4] 13731 1727203860.85930: sending task result for task 028d2410-947f-82dc-c122-0000000006a4 13731 1727203860.86134: done sending task result for task 028d2410-947f-82dc-c122-0000000006a4 13731 1727203860.86137: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203860.86243: no more pending results, returning what we have 13731 1727203860.86247: results queue empty 13731 1727203860.86248: checking for any_errors_fatal 13731 1727203860.86257: done checking for any_errors_fatal 13731 1727203860.86258: checking for max_fail_percentage 13731 1727203860.86260: done checking for max_fail_percentage 13731 1727203860.86261: checking to see if all hosts have failed and the running result is not ok 13731 1727203860.86262: done checking to see if all hosts have failed 13731 1727203860.86262: getting the remaining hosts for this loop 13731 1727203860.86264: done getting the remaining hosts for this loop 13731 1727203860.86267: getting the next task for host managed-node3 13731 1727203860.86274: done getting next task for host managed-node3 13731 1727203860.86287: ^ task is: TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13731 1727203860.86294: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=22, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203860.86420: getting variables 13731 1727203860.86422: in VariableManager get_vars() 13731 1727203860.86458: Calling all_inventory to load vars for managed-node3 13731 1727203860.86461: Calling groups_inventory to load vars for managed-node3 13731 1727203860.86464: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203860.86472: Calling all_plugins_play to load vars for managed-node3 13731 1727203860.86637: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203860.86644: Calling groups_plugins_play to load vars for managed-node3 13731 1727203860.88497: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203860.90294: done with get_vars() 13731 1727203860.90323: done getting variables 13731 1727203860.90388: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show stderr messages for the network_connections] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:177 Tuesday 24 September 2024 14:51:00 -0400 (0:00:00.069) 0:00:39.135 ***** 13731 1727203860.90426: entering _queue_task() for managed-node3/debug 13731 1727203860.90804: worker is 1 (out of 1 available) 13731 1727203860.90815: exiting _queue_task() for managed-node3/debug 13731 1727203860.90827: done queuing things up, now waiting for results queue to drain 13731 1727203860.90829: waiting for pending results... 13731 1727203860.91180: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13731 1727203860.91324: in run() - task 028d2410-947f-82dc-c122-0000000006a5 13731 1727203860.91338: variable 'ansible_search_path' from source: unknown 13731 1727203860.91384: variable 'ansible_search_path' from source: unknown 13731 1727203860.91388: calling self._execute() 13731 1727203860.91510: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203860.91516: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203860.91527: variable 'omit' from source: magic vars 13731 1727203860.91946: variable 'ansible_distribution_major_version' from source: facts 13731 1727203860.92039: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203860.92042: variable 'omit' from source: magic vars 13731 1727203860.92061: variable 'omit' from source: magic vars 13731 1727203860.92099: variable 'omit' from source: magic vars 13731 1727203860.92149: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203860.92185: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203860.92206: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203860.92234: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203860.92246: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203860.92279: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203860.92283: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203860.92285: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203860.92401: Set connection var ansible_pipelining to False 13731 1727203860.92404: Set connection var ansible_shell_type to sh 13731 1727203860.92477: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203860.92481: Set connection var ansible_connection to ssh 13731 1727203860.92483: Set connection var ansible_shell_executable to /bin/sh 13731 1727203860.92485: Set connection var ansible_timeout to 10 13731 1727203860.92487: variable 'ansible_shell_executable' from source: unknown 13731 1727203860.92489: variable 'ansible_connection' from source: unknown 13731 1727203860.92492: variable 'ansible_module_compression' from source: unknown 13731 1727203860.92494: variable 'ansible_shell_type' from source: unknown 13731 1727203860.92496: variable 'ansible_shell_executable' from source: unknown 13731 1727203860.92498: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203860.92500: variable 'ansible_pipelining' from source: unknown 13731 1727203860.92502: variable 'ansible_timeout' from source: unknown 13731 1727203860.92504: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203860.92630: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203860.92692: variable 'omit' from source: magic vars 13731 1727203860.92695: starting attempt loop 13731 1727203860.92698: running the handler 13731 1727203860.92800: variable '__network_connections_result' from source: set_fact 13731 1727203860.92849: handler run complete 13731 1727203860.92867: attempt loop complete, returning result 13731 1727203860.92879: _execute() done 13731 1727203860.92882: dumping result to json 13731 1727203860.92885: done dumping result, returning 13731 1727203860.92896: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections [028d2410-947f-82dc-c122-0000000006a5] 13731 1727203860.92901: sending task result for task 028d2410-947f-82dc-c122-0000000006a5 ok: [managed-node3] => { "__network_connections_result.stderr_lines": [ "" ] } 13731 1727203860.93065: done sending task result for task 028d2410-947f-82dc-c122-0000000006a5 13731 1727203860.93069: WORKER PROCESS EXITING 13731 1727203860.93103: no more pending results, returning what we have 13731 1727203860.93108: results queue empty 13731 1727203860.93109: checking for any_errors_fatal 13731 1727203860.93115: done checking for any_errors_fatal 13731 1727203860.93116: checking for max_fail_percentage 13731 1727203860.93118: done checking for max_fail_percentage 13731 1727203860.93119: checking to see if all hosts have failed and the running result is not ok 13731 1727203860.93120: done checking to see if all hosts have failed 13731 1727203860.93120: getting the remaining hosts for this loop 13731 1727203860.93122: done getting the remaining hosts for this loop 13731 1727203860.93127: getting the next task for host managed-node3 13731 1727203860.93135: done getting next task for host managed-node3 13731 1727203860.93140: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13731 1727203860.93146: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=23, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203860.93158: getting variables 13731 1727203860.93160: in VariableManager get_vars() 13731 1727203860.93202: Calling all_inventory to load vars for managed-node3 13731 1727203860.93205: Calling groups_inventory to load vars for managed-node3 13731 1727203860.93208: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203860.93218: Calling all_plugins_play to load vars for managed-node3 13731 1727203860.93221: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203860.93224: Calling groups_plugins_play to load vars for managed-node3 13731 1727203860.95462: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203860.97568: done with get_vars() 13731 1727203860.97601: done getting variables 13731 1727203860.97664: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_connections] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:181 Tuesday 24 September 2024 14:51:00 -0400 (0:00:00.072) 0:00:39.208 ***** 13731 1727203860.97705: entering _queue_task() for managed-node3/debug 13731 1727203860.98401: worker is 1 (out of 1 available) 13731 1727203860.98414: exiting _queue_task() for managed-node3/debug 13731 1727203860.98425: done queuing things up, now waiting for results queue to drain 13731 1727203860.98427: waiting for pending results... 13731 1727203860.98846: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13731 1727203860.98964: in run() - task 028d2410-947f-82dc-c122-0000000006a6 13731 1727203860.98978: variable 'ansible_search_path' from source: unknown 13731 1727203860.98982: variable 'ansible_search_path' from source: unknown 13731 1727203860.99019: calling self._execute() 13731 1727203860.99122: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203860.99128: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203860.99158: variable 'omit' from source: magic vars 13731 1727203860.99594: variable 'ansible_distribution_major_version' from source: facts 13731 1727203860.99597: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203860.99600: variable 'omit' from source: magic vars 13731 1727203860.99653: variable 'omit' from source: magic vars 13731 1727203860.99702: variable 'omit' from source: magic vars 13731 1727203860.99739: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203860.99775: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203860.99917: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203860.99921: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203860.99923: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203860.99925: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203860.99928: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203860.99931: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203860.99983: Set connection var ansible_pipelining to False 13731 1727203860.99989: Set connection var ansible_shell_type to sh 13731 1727203860.99995: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203861.00000: Set connection var ansible_connection to ssh 13731 1727203861.00012: Set connection var ansible_shell_executable to /bin/sh 13731 1727203861.00022: Set connection var ansible_timeout to 10 13731 1727203861.00046: variable 'ansible_shell_executable' from source: unknown 13731 1727203861.00049: variable 'ansible_connection' from source: unknown 13731 1727203861.00052: variable 'ansible_module_compression' from source: unknown 13731 1727203861.00055: variable 'ansible_shell_type' from source: unknown 13731 1727203861.00057: variable 'ansible_shell_executable' from source: unknown 13731 1727203861.00062: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203861.00064: variable 'ansible_pipelining' from source: unknown 13731 1727203861.00066: variable 'ansible_timeout' from source: unknown 13731 1727203861.00068: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203861.00230: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203861.00245: variable 'omit' from source: magic vars 13731 1727203861.00251: starting attempt loop 13731 1727203861.00255: running the handler 13731 1727203861.00306: variable '__network_connections_result' from source: set_fact 13731 1727203861.00462: variable '__network_connections_result' from source: set_fact 13731 1727203861.00528: handler run complete 13731 1727203861.00565: attempt loop complete, returning result 13731 1727203861.00572: _execute() done 13731 1727203861.00574: dumping result to json 13731 1727203861.00579: done dumping result, returning 13731 1727203861.00585: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections [028d2410-947f-82dc-c122-0000000006a6] 13731 1727203861.00590: sending task result for task 028d2410-947f-82dc-c122-0000000006a6 ok: [managed-node3] => { "__network_connections_result": { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "name": "bond0.1", "persistent_state": "absent", "state": "down" }, { "name": "bond0.0", "persistent_state": "absent", "state": "down" }, { "name": "bond0", "persistent_state": "absent", "state": "down" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true, "failed": false, "stderr": "\n", "stderr_lines": [ "" ] } } 13731 1727203861.00892: no more pending results, returning what we have 13731 1727203861.00896: results queue empty 13731 1727203861.00897: checking for any_errors_fatal 13731 1727203861.00902: done checking for any_errors_fatal 13731 1727203861.00902: checking for max_fail_percentage 13731 1727203861.00904: done checking for max_fail_percentage 13731 1727203861.00904: checking to see if all hosts have failed and the running result is not ok 13731 1727203861.00905: done checking to see if all hosts have failed 13731 1727203861.00906: getting the remaining hosts for this loop 13731 1727203861.00908: done getting the remaining hosts for this loop 13731 1727203861.00911: getting the next task for host managed-node3 13731 1727203861.00917: done getting next task for host managed-node3 13731 1727203861.00921: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13731 1727203861.00925: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=24, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203861.01003: getting variables 13731 1727203861.01005: in VariableManager get_vars() 13731 1727203861.01041: Calling all_inventory to load vars for managed-node3 13731 1727203861.01044: Calling groups_inventory to load vars for managed-node3 13731 1727203861.01046: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203861.01065: Calling all_plugins_play to load vars for managed-node3 13731 1727203861.01068: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203861.01071: Calling groups_plugins_play to load vars for managed-node3 13731 1727203861.01085: done sending task result for task 028d2410-947f-82dc-c122-0000000006a6 13731 1727203861.01088: WORKER PROCESS EXITING 13731 1727203861.02413: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203861.03996: done with get_vars() 13731 1727203861.04033: done getting variables 13731 1727203861.04099: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_state] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:186 Tuesday 24 September 2024 14:51:01 -0400 (0:00:00.064) 0:00:39.272 ***** 13731 1727203861.04140: entering _queue_task() for managed-node3/debug 13731 1727203861.04591: worker is 1 (out of 1 available) 13731 1727203861.04607: exiting _queue_task() for managed-node3/debug 13731 1727203861.04619: done queuing things up, now waiting for results queue to drain 13731 1727203861.04621: waiting for pending results... 13731 1727203861.04890: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13731 1727203861.05069: in run() - task 028d2410-947f-82dc-c122-0000000006a7 13731 1727203861.05094: variable 'ansible_search_path' from source: unknown 13731 1727203861.05103: variable 'ansible_search_path' from source: unknown 13731 1727203861.05149: calling self._execute() 13731 1727203861.05258: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203861.05334: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203861.05338: variable 'omit' from source: magic vars 13731 1727203861.05698: variable 'ansible_distribution_major_version' from source: facts 13731 1727203861.05717: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203861.05846: variable 'network_state' from source: role '' defaults 13731 1727203861.05868: Evaluated conditional (network_state != {}): False 13731 1727203861.05882: when evaluation is False, skipping this task 13731 1727203861.05892: _execute() done 13731 1727203861.05899: dumping result to json 13731 1727203861.05907: done dumping result, returning 13731 1727203861.05919: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state [028d2410-947f-82dc-c122-0000000006a7] 13731 1727203861.05982: sending task result for task 028d2410-947f-82dc-c122-0000000006a7 13731 1727203861.06061: done sending task result for task 028d2410-947f-82dc-c122-0000000006a7 13731 1727203861.06066: WORKER PROCESS EXITING skipping: [managed-node3] => { "false_condition": "network_state != {}" } 13731 1727203861.06123: no more pending results, returning what we have 13731 1727203861.06127: results queue empty 13731 1727203861.06128: checking for any_errors_fatal 13731 1727203861.06138: done checking for any_errors_fatal 13731 1727203861.06139: checking for max_fail_percentage 13731 1727203861.06140: done checking for max_fail_percentage 13731 1727203861.06141: checking to see if all hosts have failed and the running result is not ok 13731 1727203861.06142: done checking to see if all hosts have failed 13731 1727203861.06142: getting the remaining hosts for this loop 13731 1727203861.06144: done getting the remaining hosts for this loop 13731 1727203861.06148: getting the next task for host managed-node3 13731 1727203861.06155: done getting next task for host managed-node3 13731 1727203861.06159: ^ task is: TASK: fedora.linux_system_roles.network : Re-test connectivity 13731 1727203861.06164: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=25, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203861.06184: getting variables 13731 1727203861.06186: in VariableManager get_vars() 13731 1727203861.06338: Calling all_inventory to load vars for managed-node3 13731 1727203861.06341: Calling groups_inventory to load vars for managed-node3 13731 1727203861.06343: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203861.06352: Calling all_plugins_play to load vars for managed-node3 13731 1727203861.06354: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203861.06356: Calling groups_plugins_play to load vars for managed-node3 13731 1727203861.07942: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203861.09537: done with get_vars() 13731 1727203861.09572: done getting variables TASK [fedora.linux_system_roles.network : Re-test connectivity] **************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:192 Tuesday 24 September 2024 14:51:01 -0400 (0:00:00.055) 0:00:39.327 ***** 13731 1727203861.09683: entering _queue_task() for managed-node3/ping 13731 1727203861.10069: worker is 1 (out of 1 available) 13731 1727203861.10289: exiting _queue_task() for managed-node3/ping 13731 1727203861.10302: done queuing things up, now waiting for results queue to drain 13731 1727203861.10304: waiting for pending results... 13731 1727203861.10545: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Re-test connectivity 13731 1727203861.10615: in run() - task 028d2410-947f-82dc-c122-0000000006a8 13731 1727203861.10648: variable 'ansible_search_path' from source: unknown 13731 1727203861.10658: variable 'ansible_search_path' from source: unknown 13731 1727203861.10707: calling self._execute() 13731 1727203861.10863: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203861.10867: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203861.10870: variable 'omit' from source: magic vars 13731 1727203861.11292: variable 'ansible_distribution_major_version' from source: facts 13731 1727203861.11314: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203861.11380: variable 'omit' from source: magic vars 13731 1727203861.11418: variable 'omit' from source: magic vars 13731 1727203861.11464: variable 'omit' from source: magic vars 13731 1727203861.11523: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203861.11569: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203861.11598: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203861.11631: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203861.11649: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203861.11692: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203861.11728: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203861.11734: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203861.11827: Set connection var ansible_pipelining to False 13731 1727203861.11847: Set connection var ansible_shell_type to sh 13731 1727203861.11946: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203861.11950: Set connection var ansible_connection to ssh 13731 1727203861.11952: Set connection var ansible_shell_executable to /bin/sh 13731 1727203861.11954: Set connection var ansible_timeout to 10 13731 1727203861.11956: variable 'ansible_shell_executable' from source: unknown 13731 1727203861.11958: variable 'ansible_connection' from source: unknown 13731 1727203861.11962: variable 'ansible_module_compression' from source: unknown 13731 1727203861.11964: variable 'ansible_shell_type' from source: unknown 13731 1727203861.11967: variable 'ansible_shell_executable' from source: unknown 13731 1727203861.11969: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203861.11971: variable 'ansible_pipelining' from source: unknown 13731 1727203861.11972: variable 'ansible_timeout' from source: unknown 13731 1727203861.11975: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203861.12271: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203861.12278: variable 'omit' from source: magic vars 13731 1727203861.12281: starting attempt loop 13731 1727203861.12283: running the handler 13731 1727203861.12285: _low_level_execute_command(): starting 13731 1727203861.12287: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203861.13055: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203861.13088: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.13153: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.13208: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.13226: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.13258: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.13328: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.15188: stdout chunk (state=3): >>>/root <<< 13731 1727203861.15212: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.15287: stderr chunk (state=3): >>><<< 13731 1727203861.15468: stdout chunk (state=3): >>><<< 13731 1727203861.15473: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203861.15479: _low_level_execute_command(): starting 13731 1727203861.15483: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913 `" && echo ansible-tmp-1727203861.1539662-16559-185428939817913="` echo /root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913 `" ) && sleep 0' 13731 1727203861.17534: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.17542: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203861.17553: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.17630: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.17962: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.17967: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.18216: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.20138: stdout chunk (state=3): >>>ansible-tmp-1727203861.1539662-16559-185428939817913=/root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913 <<< 13731 1727203861.20305: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.20308: stdout chunk (state=3): >>><<< 13731 1727203861.20310: stderr chunk (state=3): >>><<< 13731 1727203861.20316: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203861.1539662-16559-185428939817913=/root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203861.20524: variable 'ansible_module_compression' from source: unknown 13731 1727203861.20527: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.ping-ZIP_DEFLATED 13731 1727203861.20529: variable 'ansible_facts' from source: unknown 13731 1727203861.20691: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/AnsiballZ_ping.py 13731 1727203861.21120: Sending initial data 13731 1727203861.21123: Sent initial data (153 bytes) 13731 1727203861.22281: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203861.22286: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203861.22289: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.22291: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203861.22293: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203861.22295: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203861.22298: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.22436: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.22472: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.24020: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203861.24050: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203861.24094: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp0tu5cst1 /root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/AnsiballZ_ping.py <<< 13731 1727203861.24098: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/AnsiballZ_ping.py" <<< 13731 1727203861.24131: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp0tu5cst1" to remote "/root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/AnsiballZ_ping.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/AnsiballZ_ping.py" <<< 13731 1727203861.25513: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.25567: stderr chunk (state=3): >>><<< 13731 1727203861.25582: stdout chunk (state=3): >>><<< 13731 1727203861.25593: done transferring module to remote 13731 1727203861.25683: _low_level_execute_command(): starting 13731 1727203861.25688: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/ /root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/AnsiballZ_ping.py && sleep 0' 13731 1727203861.27066: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203861.27230: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203861.27340: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.27344: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.27367: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.27381: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.27395: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.27572: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.29370: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.29437: stdout chunk (state=3): >>><<< 13731 1727203861.29440: stderr chunk (state=3): >>><<< 13731 1727203861.29585: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203861.29589: _low_level_execute_command(): starting 13731 1727203861.29591: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/AnsiballZ_ping.py && sleep 0' 13731 1727203861.30765: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.30950: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.30979: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.31252: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.45920: stdout chunk (state=3): >>> {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} <<< 13731 1727203861.47194: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203861.47200: stderr chunk (state=3): >>><<< 13731 1727203861.47208: stdout chunk (state=3): >>><<< 13731 1727203861.47233: _low_level_execute_command() done: rc=0, stdout= {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203861.47354: done with _execute_module (ping, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ping', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203861.47357: _low_level_execute_command(): starting 13731 1727203861.47363: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203861.1539662-16559-185428939817913/ > /dev/null 2>&1 && sleep 0' 13731 1727203861.47933: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203861.47994: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.48056: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.48080: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.48102: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.48154: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.50065: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.50070: stderr chunk (state=3): >>><<< 13731 1727203861.50072: stdout chunk (state=3): >>><<< 13731 1727203861.50182: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203861.50193: handler run complete 13731 1727203861.50195: attempt loop complete, returning result 13731 1727203861.50198: _execute() done 13731 1727203861.50200: dumping result to json 13731 1727203861.50202: done dumping result, returning 13731 1727203861.50204: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Re-test connectivity [028d2410-947f-82dc-c122-0000000006a8] 13731 1727203861.50206: sending task result for task 028d2410-947f-82dc-c122-0000000006a8 ok: [managed-node3] => { "changed": false, "ping": "pong" } 13731 1727203861.50351: no more pending results, returning what we have 13731 1727203861.50356: results queue empty 13731 1727203861.50357: checking for any_errors_fatal 13731 1727203861.50367: done checking for any_errors_fatal 13731 1727203861.50368: checking for max_fail_percentage 13731 1727203861.50370: done checking for max_fail_percentage 13731 1727203861.50371: checking to see if all hosts have failed and the running result is not ok 13731 1727203861.50371: done checking to see if all hosts have failed 13731 1727203861.50372: getting the remaining hosts for this loop 13731 1727203861.50374: done getting the remaining hosts for this loop 13731 1727203861.50482: getting the next task for host managed-node3 13731 1727203861.50495: done getting next task for host managed-node3 13731 1727203861.50498: ^ task is: TASK: meta (role_complete) 13731 1727203861.50504: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203861.50517: getting variables 13731 1727203861.50519: in VariableManager get_vars() 13731 1727203861.50564: Calling all_inventory to load vars for managed-node3 13731 1727203861.50567: Calling groups_inventory to load vars for managed-node3 13731 1727203861.50570: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203861.50700: Calling all_plugins_play to load vars for managed-node3 13731 1727203861.50704: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203861.50709: Calling groups_plugins_play to load vars for managed-node3 13731 1727203861.51342: done sending task result for task 028d2410-947f-82dc-c122-0000000006a8 13731 1727203861.51347: WORKER PROCESS EXITING 13731 1727203861.52407: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203861.54458: done with get_vars() 13731 1727203861.54484: done getting variables 13731 1727203861.54567: done queuing things up, now waiting for results queue to drain 13731 1727203861.54569: results queue empty 13731 1727203861.54570: checking for any_errors_fatal 13731 1727203861.54573: done checking for any_errors_fatal 13731 1727203861.54573: checking for max_fail_percentage 13731 1727203861.54574: done checking for max_fail_percentage 13731 1727203861.54649: checking to see if all hosts have failed and the running result is not ok 13731 1727203861.54652: done checking to see if all hosts have failed 13731 1727203861.54653: getting the remaining hosts for this loop 13731 1727203861.54654: done getting the remaining hosts for this loop 13731 1727203861.54658: getting the next task for host managed-node3 13731 1727203861.54666: done getting next task for host managed-node3 13731 1727203861.54668: ^ task is: TASK: Delete the device '{{ controller_device }}' 13731 1727203861.54671: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203861.54674: getting variables 13731 1727203861.54676: in VariableManager get_vars() 13731 1727203861.54691: Calling all_inventory to load vars for managed-node3 13731 1727203861.54693: Calling groups_inventory to load vars for managed-node3 13731 1727203861.54695: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203861.54701: Calling all_plugins_play to load vars for managed-node3 13731 1727203861.54703: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203861.54706: Calling groups_plugins_play to load vars for managed-node3 13731 1727203861.55905: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203861.57482: done with get_vars() 13731 1727203861.57506: done getting variables 13731 1727203861.57550: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203861.57689: variable 'controller_device' from source: play vars TASK [Delete the device 'nm-bond'] ********************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml:22 Tuesday 24 September 2024 14:51:01 -0400 (0:00:00.480) 0:00:39.808 ***** 13731 1727203861.57721: entering _queue_task() for managed-node3/command 13731 1727203861.58107: worker is 1 (out of 1 available) 13731 1727203861.58189: exiting _queue_task() for managed-node3/command 13731 1727203861.58200: done queuing things up, now waiting for results queue to drain 13731 1727203861.58202: waiting for pending results... 13731 1727203861.58444: running TaskExecutor() for managed-node3/TASK: Delete the device 'nm-bond' 13731 1727203861.58673: in run() - task 028d2410-947f-82dc-c122-0000000006d8 13731 1727203861.58679: variable 'ansible_search_path' from source: unknown 13731 1727203861.58682: variable 'ansible_search_path' from source: unknown 13731 1727203861.58684: calling self._execute() 13731 1727203861.58755: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203861.58770: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203861.58796: variable 'omit' from source: magic vars 13731 1727203861.59172: variable 'ansible_distribution_major_version' from source: facts 13731 1727203861.59192: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203861.59215: variable 'omit' from source: magic vars 13731 1727203861.59282: variable 'omit' from source: magic vars 13731 1727203861.59344: variable 'controller_device' from source: play vars 13731 1727203861.59368: variable 'omit' from source: magic vars 13731 1727203861.59412: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203861.59461: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203861.59492: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203861.59514: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203861.59531: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203861.59583: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203861.59586: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203861.59591: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203861.59707: Set connection var ansible_pipelining to False 13731 1727203861.59718: Set connection var ansible_shell_type to sh 13731 1727203861.59728: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203861.59738: Set connection var ansible_connection to ssh 13731 1727203861.59748: Set connection var ansible_shell_executable to /bin/sh 13731 1727203861.59766: Set connection var ansible_timeout to 10 13731 1727203861.59800: variable 'ansible_shell_executable' from source: unknown 13731 1727203861.59866: variable 'ansible_connection' from source: unknown 13731 1727203861.59869: variable 'ansible_module_compression' from source: unknown 13731 1727203861.59871: variable 'ansible_shell_type' from source: unknown 13731 1727203861.59873: variable 'ansible_shell_executable' from source: unknown 13731 1727203861.59874: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203861.59877: variable 'ansible_pipelining' from source: unknown 13731 1727203861.59880: variable 'ansible_timeout' from source: unknown 13731 1727203861.59882: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203861.59995: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203861.60010: variable 'omit' from source: magic vars 13731 1727203861.60027: starting attempt loop 13731 1727203861.60034: running the handler 13731 1727203861.60056: _low_level_execute_command(): starting 13731 1727203861.60073: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203861.60952: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.60974: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.61045: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.61073: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.61113: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.61189: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.62898: stdout chunk (state=3): >>>/root <<< 13731 1727203861.63057: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.63064: stdout chunk (state=3): >>><<< 13731 1727203861.63067: stderr chunk (state=3): >>><<< 13731 1727203861.63197: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203861.63202: _low_level_execute_command(): starting 13731 1727203861.63205: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063 `" && echo ansible-tmp-1727203861.6309953-16592-26226501906063="` echo /root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063 `" ) && sleep 0' 13731 1727203861.63997: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203861.64019: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203861.64035: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.64091: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.64169: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.64190: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.64222: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.64294: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.66310: stdout chunk (state=3): >>>ansible-tmp-1727203861.6309953-16592-26226501906063=/root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063 <<< 13731 1727203861.66461: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.66486: stdout chunk (state=3): >>><<< 13731 1727203861.66499: stderr chunk (state=3): >>><<< 13731 1727203861.66522: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203861.6309953-16592-26226501906063=/root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203861.66581: variable 'ansible_module_compression' from source: unknown 13731 1727203861.66631: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203861.66677: variable 'ansible_facts' from source: unknown 13731 1727203861.66882: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/AnsiballZ_command.py 13731 1727203861.66930: Sending initial data 13731 1727203861.66939: Sent initial data (155 bytes) 13731 1727203861.67666: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203861.67783: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.67804: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.67807: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.67874: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.67883: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.67991: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.68064: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.69715: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203861.69796: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203861.69823: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp11za03hs /root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/AnsiballZ_command.py <<< 13731 1727203861.69826: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/AnsiballZ_command.py" <<< 13731 1727203861.69871: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp11za03hs" to remote "/root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/AnsiballZ_command.py" <<< 13731 1727203861.70649: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.70695: stderr chunk (state=3): >>><<< 13731 1727203861.70768: stdout chunk (state=3): >>><<< 13731 1727203861.70784: done transferring module to remote 13731 1727203861.70798: _low_level_execute_command(): starting 13731 1727203861.70806: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/ /root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/AnsiballZ_command.py && sleep 0' 13731 1727203861.71518: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203861.71591: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203861.71663: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203861.71681: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.71718: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.71739: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.71783: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.71869: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.73757: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.73774: stderr chunk (state=3): >>><<< 13731 1727203861.73800: stdout chunk (state=3): >>><<< 13731 1727203861.73824: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203861.73832: _low_level_execute_command(): starting 13731 1727203861.73842: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/AnsiballZ_command.py && sleep 0' 13731 1727203861.74515: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203861.74547: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203861.74561: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.74579: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203861.74595: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203861.74604: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203861.74650: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.74722: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203861.74753: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.74819: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.91332: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "Cannot find device \"nm-bond\"", "rc": 1, "cmd": ["ip", "link", "del", "nm-bond"], "start": "2024-09-24 14:51:01.904869", "end": "2024-09-24 14:51:01.912372", "delta": "0:00:00.007503", "failed": true, "msg": "non-zero return code", "invocation": {"module_args": {"_raw_params": "ip link del nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203861.92884: stderr chunk (state=3): >>>debug2: Received exit status from master 1 Shared connection to 10.31.47.22 closed. <<< 13731 1727203861.92888: stdout chunk (state=3): >>><<< 13731 1727203861.92890: stderr chunk (state=3): >>><<< 13731 1727203861.92974: _low_level_execute_command() done: rc=1, stdout= {"changed": true, "stdout": "", "stderr": "Cannot find device \"nm-bond\"", "rc": 1, "cmd": ["ip", "link", "del", "nm-bond"], "start": "2024-09-24 14:51:01.904869", "end": "2024-09-24 14:51:01.912372", "delta": "0:00:00.007503", "failed": true, "msg": "non-zero return code", "invocation": {"module_args": {"_raw_params": "ip link del nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 1 Shared connection to 10.31.47.22 closed. 13731 1727203861.92981: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip link del nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203861.92984: _low_level_execute_command(): starting 13731 1727203861.92987: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203861.6309953-16592-26226501906063/ > /dev/null 2>&1 && sleep 0' 13731 1727203861.93581: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203861.93590: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203861.93602: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.93622: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203861.93628: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203861.93635: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203861.93645: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.93662: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203861.93671: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203861.93674: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203861.93683: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203861.93730: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203861.93734: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203861.93736: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203861.93738: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203861.93740: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203861.93844: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203861.93847: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203861.93880: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203861.95797: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203861.95801: stdout chunk (state=3): >>><<< 13731 1727203861.95803: stderr chunk (state=3): >>><<< 13731 1727203861.95822: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203861.95834: handler run complete 13731 1727203861.95867: Evaluated conditional (False): False 13731 1727203861.95984: Evaluated conditional (False): False 13731 1727203861.95997: attempt loop complete, returning result 13731 1727203861.96000: _execute() done 13731 1727203861.96002: dumping result to json 13731 1727203861.96005: done dumping result, returning 13731 1727203861.96007: done running TaskExecutor() for managed-node3/TASK: Delete the device 'nm-bond' [028d2410-947f-82dc-c122-0000000006d8] 13731 1727203861.96009: sending task result for task 028d2410-947f-82dc-c122-0000000006d8 13731 1727203861.96085: done sending task result for task 028d2410-947f-82dc-c122-0000000006d8 13731 1727203861.96089: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": [ "ip", "link", "del", "nm-bond" ], "delta": "0:00:00.007503", "end": "2024-09-24 14:51:01.912372", "failed_when_result": false, "rc": 1, "start": "2024-09-24 14:51:01.904869" } STDERR: Cannot find device "nm-bond" MSG: non-zero return code 13731 1727203861.96164: no more pending results, returning what we have 13731 1727203861.96168: results queue empty 13731 1727203861.96169: checking for any_errors_fatal 13731 1727203861.96172: done checking for any_errors_fatal 13731 1727203861.96172: checking for max_fail_percentage 13731 1727203861.96174: done checking for max_fail_percentage 13731 1727203861.96178: checking to see if all hosts have failed and the running result is not ok 13731 1727203861.96179: done checking to see if all hosts have failed 13731 1727203861.96179: getting the remaining hosts for this loop 13731 1727203861.96181: done getting the remaining hosts for this loop 13731 1727203861.96184: getting the next task for host managed-node3 13731 1727203861.96194: done getting next task for host managed-node3 13731 1727203861.96197: ^ task is: TASK: Remove test interfaces 13731 1727203861.96202: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203861.96327: getting variables 13731 1727203861.96329: in VariableManager get_vars() 13731 1727203861.96366: Calling all_inventory to load vars for managed-node3 13731 1727203861.96369: Calling groups_inventory to load vars for managed-node3 13731 1727203861.96371: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203861.96390: Calling all_plugins_play to load vars for managed-node3 13731 1727203861.96393: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203861.96396: Calling groups_plugins_play to load vars for managed-node3 13731 1727203861.98027: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203861.99968: done with get_vars() 13731 1727203862.00008: done getting variables 13731 1727203862.00085: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Remove test interfaces] ************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml:3 Tuesday 24 September 2024 14:51:02 -0400 (0:00:00.423) 0:00:40.232 ***** 13731 1727203862.00122: entering _queue_task() for managed-node3/shell 13731 1727203862.00528: worker is 1 (out of 1 available) 13731 1727203862.00542: exiting _queue_task() for managed-node3/shell 13731 1727203862.00555: done queuing things up, now waiting for results queue to drain 13731 1727203862.00557: waiting for pending results... 13731 1727203862.00999: running TaskExecutor() for managed-node3/TASK: Remove test interfaces 13731 1727203862.01038: in run() - task 028d2410-947f-82dc-c122-0000000006de 13731 1727203862.01064: variable 'ansible_search_path' from source: unknown 13731 1727203862.01073: variable 'ansible_search_path' from source: unknown 13731 1727203862.01124: calling self._execute() 13731 1727203862.01235: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203862.01247: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203862.01264: variable 'omit' from source: magic vars 13731 1727203862.01685: variable 'ansible_distribution_major_version' from source: facts 13731 1727203862.01705: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203862.01717: variable 'omit' from source: magic vars 13731 1727203862.01784: variable 'omit' from source: magic vars 13731 1727203862.01967: variable 'dhcp_interface1' from source: play vars 13731 1727203862.01982: variable 'dhcp_interface2' from source: play vars 13731 1727203862.02005: variable 'omit' from source: magic vars 13731 1727203862.02052: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203862.02222: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203862.02399: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203862.02402: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203862.02507: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203862.02510: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203862.02513: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203862.02515: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203862.02622: Set connection var ansible_pipelining to False 13731 1727203862.02634: Set connection var ansible_shell_type to sh 13731 1727203862.02642: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203862.02651: Set connection var ansible_connection to ssh 13731 1727203862.02665: Set connection var ansible_shell_executable to /bin/sh 13731 1727203862.02678: Set connection var ansible_timeout to 10 13731 1727203862.02706: variable 'ansible_shell_executable' from source: unknown 13731 1727203862.02715: variable 'ansible_connection' from source: unknown 13731 1727203862.02740: variable 'ansible_module_compression' from source: unknown 13731 1727203862.02743: variable 'ansible_shell_type' from source: unknown 13731 1727203862.02746: variable 'ansible_shell_executable' from source: unknown 13731 1727203862.02833: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203862.02836: variable 'ansible_pipelining' from source: unknown 13731 1727203862.02839: variable 'ansible_timeout' from source: unknown 13731 1727203862.02841: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203862.03052: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203862.03056: variable 'omit' from source: magic vars 13731 1727203862.03063: starting attempt loop 13731 1727203862.03068: running the handler 13731 1727203862.03071: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203862.03073: _low_level_execute_command(): starting 13731 1727203862.03077: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203862.03782: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203862.03807: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203862.03825: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203862.03850: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203862.03927: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.03983: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.04008: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.04064: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.04094: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.05972: stdout chunk (state=3): >>>/root <<< 13731 1727203862.06386: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.06389: stdout chunk (state=3): >>><<< 13731 1727203862.06392: stderr chunk (state=3): >>><<< 13731 1727203862.06395: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203862.06397: _low_level_execute_command(): starting 13731 1727203862.06400: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953 `" && echo ansible-tmp-1727203862.0634055-16611-230035602749953="` echo /root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953 `" ) && sleep 0' 13731 1727203862.08686: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203862.08690: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203862.08694: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203862.08696: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203862.08968: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.09004: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.09461: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.11526: stdout chunk (state=3): >>>ansible-tmp-1727203862.0634055-16611-230035602749953=/root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953 <<< 13731 1727203862.11529: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.11691: stderr chunk (state=3): >>><<< 13731 1727203862.11695: stdout chunk (state=3): >>><<< 13731 1727203862.11717: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203862.0634055-16611-230035602749953=/root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203862.11837: variable 'ansible_module_compression' from source: unknown 13731 1727203862.12054: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203862.12057: variable 'ansible_facts' from source: unknown 13731 1727203862.12333: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/AnsiballZ_command.py 13731 1727203862.13093: Sending initial data 13731 1727203862.13097: Sent initial data (156 bytes) 13731 1727203862.14631: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.14672: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.16262: stderr chunk (state=3): >>>debug2: Remote version: 3 <<< 13731 1727203862.16403: stderr chunk (state=3): >>>debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203862.16435: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203862.16502: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/AnsiballZ_command.py" <<< 13731 1727203862.16512: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpjcnx33hw /root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/AnsiballZ_command.py <<< 13731 1727203862.16531: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpjcnx33hw" to remote "/root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/AnsiballZ_command.py" <<< 13731 1727203862.17799: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.17892: stderr chunk (state=3): >>><<< 13731 1727203862.17999: stdout chunk (state=3): >>><<< 13731 1727203862.18003: done transferring module to remote 13731 1727203862.18005: _low_level_execute_command(): starting 13731 1727203862.18008: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/ /root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/AnsiballZ_command.py && sleep 0' 13731 1727203862.18909: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.18995: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.19093: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.20947: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.20957: stdout chunk (state=3): >>><<< 13731 1727203862.21140: stderr chunk (state=3): >>><<< 13731 1727203862.21147: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203862.21149: _low_level_execute_command(): starting 13731 1727203862.21152: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/AnsiballZ_command.py && sleep 0' 13731 1727203862.21920: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203862.21925: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.21927: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203862.21942: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.21995: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.22017: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.22021: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.22112: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.40928: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ rc=0\n+ ip link delete test1\n+ '[' 0 '!=' 0 ']'\n+ ip link delete test2\n+ '[' 0 '!=' 0 ']'\n+ ip link delete testbr\n+ '[' 0 '!=' 0 ']'", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "start": "2024-09-24 14:51:02.372019", "end": "2024-09-24 14:51:02.407675", "delta": "0:00:00.035656", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203862.42572: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203862.42967: stdout chunk (state=3): >>><<< 13731 1727203862.42971: stderr chunk (state=3): >>><<< 13731 1727203862.42974: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ rc=0\n+ ip link delete test1\n+ '[' 0 '!=' 0 ']'\n+ ip link delete test2\n+ '[' 0 '!=' 0 ']'\n+ ip link delete testbr\n+ '[' 0 '!=' 0 ']'", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "start": "2024-09-24 14:51:02.372019", "end": "2024-09-24 14:51:02.407675", "delta": "0:00:00.035656", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203862.42979: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link test1 - error "$rc"\nfi\nip link delete test2 || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link test2 - error "$rc"\nfi\nip link delete testbr || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link testbr - error "$rc"\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203862.42987: _low_level_execute_command(): starting 13731 1727203862.42990: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203862.0634055-16611-230035602749953/ > /dev/null 2>&1 && sleep 0' 13731 1727203862.44192: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203862.44211: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203862.44291: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.44492: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.44543: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.44647: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.46514: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.46633: stderr chunk (state=3): >>><<< 13731 1727203862.46636: stdout chunk (state=3): >>><<< 13731 1727203862.46728: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203862.46732: handler run complete 13731 1727203862.46774: Evaluated conditional (False): False 13731 1727203862.46794: attempt loop complete, returning result 13731 1727203862.46797: _execute() done 13731 1727203862.46799: dumping result to json 13731 1727203862.46801: done dumping result, returning 13731 1727203862.46982: done running TaskExecutor() for managed-node3/TASK: Remove test interfaces [028d2410-947f-82dc-c122-0000000006de] 13731 1727203862.46986: sending task result for task 028d2410-947f-82dc-c122-0000000006de 13731 1727203862.47061: done sending task result for task 028d2410-947f-82dc-c122-0000000006de 13731 1727203862.47064: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "delta": "0:00:00.035656", "end": "2024-09-24 14:51:02.407675", "rc": 0, "start": "2024-09-24 14:51:02.372019" } STDERR: + exec + rc=0 + ip link delete test1 + '[' 0 '!=' 0 ']' + ip link delete test2 + '[' 0 '!=' 0 ']' + ip link delete testbr + '[' 0 '!=' 0 ']' 13731 1727203862.47423: no more pending results, returning what we have 13731 1727203862.47427: results queue empty 13731 1727203862.47428: checking for any_errors_fatal 13731 1727203862.47435: done checking for any_errors_fatal 13731 1727203862.47436: checking for max_fail_percentage 13731 1727203862.47437: done checking for max_fail_percentage 13731 1727203862.47438: checking to see if all hosts have failed and the running result is not ok 13731 1727203862.47439: done checking to see if all hosts have failed 13731 1727203862.47439: getting the remaining hosts for this loop 13731 1727203862.47441: done getting the remaining hosts for this loop 13731 1727203862.47445: getting the next task for host managed-node3 13731 1727203862.47453: done getting next task for host managed-node3 13731 1727203862.47455: ^ task is: TASK: Stop dnsmasq/radvd services 13731 1727203862.47459: ^ state is: HOST STATE: block=3, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203862.47463: getting variables 13731 1727203862.47464: in VariableManager get_vars() 13731 1727203862.47621: Calling all_inventory to load vars for managed-node3 13731 1727203862.47625: Calling groups_inventory to load vars for managed-node3 13731 1727203862.47627: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203862.47638: Calling all_plugins_play to load vars for managed-node3 13731 1727203862.47640: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203862.47643: Calling groups_plugins_play to load vars for managed-node3 13731 1727203862.50127: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203862.52038: done with get_vars() 13731 1727203862.52157: done getting variables 13731 1727203862.52328: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Stop dnsmasq/radvd services] ********************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml:23 Tuesday 24 September 2024 14:51:02 -0400 (0:00:00.522) 0:00:40.754 ***** 13731 1727203862.52395: entering _queue_task() for managed-node3/shell 13731 1727203862.53424: worker is 1 (out of 1 available) 13731 1727203862.53437: exiting _queue_task() for managed-node3/shell 13731 1727203862.53448: done queuing things up, now waiting for results queue to drain 13731 1727203862.53449: waiting for pending results... 13731 1727203862.53989: running TaskExecutor() for managed-node3/TASK: Stop dnsmasq/radvd services 13731 1727203862.54389: in run() - task 028d2410-947f-82dc-c122-0000000006df 13731 1727203862.54394: variable 'ansible_search_path' from source: unknown 13731 1727203862.54396: variable 'ansible_search_path' from source: unknown 13731 1727203862.54399: calling self._execute() 13731 1727203862.54429: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203862.54442: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203862.54461: variable 'omit' from source: magic vars 13731 1727203862.54873: variable 'ansible_distribution_major_version' from source: facts 13731 1727203862.54896: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203862.54909: variable 'omit' from source: magic vars 13731 1727203862.54973: variable 'omit' from source: magic vars 13731 1727203862.55018: variable 'omit' from source: magic vars 13731 1727203862.55099: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203862.55182: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203862.55281: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203862.55306: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203862.55323: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203862.55401: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203862.55438: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203862.55447: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203862.55579: Set connection var ansible_pipelining to False 13731 1727203862.55715: Set connection var ansible_shell_type to sh 13731 1727203862.55718: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203862.55720: Set connection var ansible_connection to ssh 13731 1727203862.55722: Set connection var ansible_shell_executable to /bin/sh 13731 1727203862.55724: Set connection var ansible_timeout to 10 13731 1727203862.55726: variable 'ansible_shell_executable' from source: unknown 13731 1727203862.55728: variable 'ansible_connection' from source: unknown 13731 1727203862.55730: variable 'ansible_module_compression' from source: unknown 13731 1727203862.55732: variable 'ansible_shell_type' from source: unknown 13731 1727203862.55734: variable 'ansible_shell_executable' from source: unknown 13731 1727203862.55736: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203862.55740: variable 'ansible_pipelining' from source: unknown 13731 1727203862.55742: variable 'ansible_timeout' from source: unknown 13731 1727203862.55743: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203862.55879: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203862.56039: variable 'omit' from source: magic vars 13731 1727203862.56042: starting attempt loop 13731 1727203862.56045: running the handler 13731 1727203862.56047: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203862.56049: _low_level_execute_command(): starting 13731 1727203862.56051: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203862.56937: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203862.57045: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.57066: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.57367: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.57799: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.59290: stdout chunk (state=3): >>>/root <<< 13731 1727203862.59441: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.59444: stdout chunk (state=3): >>><<< 13731 1727203862.59447: stderr chunk (state=3): >>><<< 13731 1727203862.59472: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203862.59749: _low_level_execute_command(): starting 13731 1727203862.59753: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607 `" && echo ansible-tmp-1727203862.5953336-16684-37030748824607="` echo /root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607 `" ) && sleep 0' 13731 1727203862.60905: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203862.60926: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.61143: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.61255: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.61258: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.63130: stdout chunk (state=3): >>>ansible-tmp-1727203862.5953336-16684-37030748824607=/root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607 <<< 13731 1727203862.63553: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.63557: stdout chunk (state=3): >>><<< 13731 1727203862.63563: stderr chunk (state=3): >>><<< 13731 1727203862.63565: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203862.5953336-16684-37030748824607=/root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203862.63584: variable 'ansible_module_compression' from source: unknown 13731 1727203862.63789: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203862.63829: variable 'ansible_facts' from source: unknown 13731 1727203862.64181: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/AnsiballZ_command.py 13731 1727203862.64451: Sending initial data 13731 1727203862.64455: Sent initial data (155 bytes) 13731 1727203862.66009: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.66037: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.66046: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.66065: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.66089: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.67636: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203862.67703: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203862.67757: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp09d3jq40 /root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/AnsiballZ_command.py <<< 13731 1727203862.67761: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/AnsiballZ_command.py" <<< 13731 1727203862.67850: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp09d3jq40" to remote "/root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/AnsiballZ_command.py" <<< 13731 1727203862.68720: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.68724: stderr chunk (state=3): >>><<< 13731 1727203862.68726: stdout chunk (state=3): >>><<< 13731 1727203862.68728: done transferring module to remote 13731 1727203862.68736: _low_level_execute_command(): starting 13731 1727203862.68742: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/ /root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/AnsiballZ_command.py && sleep 0' 13731 1727203862.69864: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.69890: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.69911: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.69971: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.72228: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.72232: stdout chunk (state=3): >>><<< 13731 1727203862.72235: stderr chunk (state=3): >>><<< 13731 1727203862.72238: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203862.72245: _low_level_execute_command(): starting 13731 1727203862.72247: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/AnsiballZ_command.py && sleep 0' 13731 1727203862.73017: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203862.73025: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203862.73105: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.73145: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.73160: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.73209: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.73381: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.91274: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ pkill -F /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.lease\n+ grep 'release 6' /etc/redhat-release\n+ systemctl is-active firewalld\ninactive", "rc": 0, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "start": "2024-09-24 14:51:02.885266", "end": "2024-09-24 14:51:02.911142", "delta": "0:00:00.025876", "msg": "", "invocation": {"module_args": {"_raw_params": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203862.92927: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203862.92931: stdout chunk (state=3): >>><<< 13731 1727203862.92934: stderr chunk (state=3): >>><<< 13731 1727203862.92936: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ pkill -F /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.lease\n+ grep 'release 6' /etc/redhat-release\n+ systemctl is-active firewalld\ninactive", "rc": 0, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "start": "2024-09-24 14:51:02.885266", "end": "2024-09-24 14:51:02.911142", "delta": "0:00:00.025876", "msg": "", "invocation": {"module_args": {"_raw_params": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203862.92978: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep \'release 6\' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service="$service"; then\n firewall-cmd --remove-service "$service"\n fi\n done\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203862.92987: _low_level_execute_command(): starting 13731 1727203862.92992: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203862.5953336-16684-37030748824607/ > /dev/null 2>&1 && sleep 0' 13731 1727203862.93686: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203862.93690: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203862.93693: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203862.93695: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203862.93698: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203862.93700: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203862.93702: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.93704: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203862.93706: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203862.93708: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203862.93709: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203862.93711: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203862.93713: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203862.93714: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203862.93716: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203862.93718: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203862.93765: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203862.93789: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203862.93796: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203862.93855: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203862.95646: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203862.95711: stderr chunk (state=3): >>><<< 13731 1727203862.95721: stdout chunk (state=3): >>><<< 13731 1727203862.95749: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203862.95768: handler run complete 13731 1727203862.95796: Evaluated conditional (False): False 13731 1727203862.95808: attempt loop complete, returning result 13731 1727203862.95814: _execute() done 13731 1727203862.95981: dumping result to json 13731 1727203862.95984: done dumping result, returning 13731 1727203862.95986: done running TaskExecutor() for managed-node3/TASK: Stop dnsmasq/radvd services [028d2410-947f-82dc-c122-0000000006df] 13731 1727203862.95988: sending task result for task 028d2410-947f-82dc-c122-0000000006df 13731 1727203862.96056: done sending task result for task 028d2410-947f-82dc-c122-0000000006df 13731 1727203862.96062: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "delta": "0:00:00.025876", "end": "2024-09-24 14:51:02.911142", "rc": 0, "start": "2024-09-24 14:51:02.885266" } STDERR: + exec + pkill -F /run/dhcp_testbr.pid + rm -rf /run/dhcp_testbr.pid + rm -rf /run/dhcp_testbr.lease + grep 'release 6' /etc/redhat-release + systemctl is-active firewalld inactive 13731 1727203862.96126: no more pending results, returning what we have 13731 1727203862.96130: results queue empty 13731 1727203862.96131: checking for any_errors_fatal 13731 1727203862.96144: done checking for any_errors_fatal 13731 1727203862.96145: checking for max_fail_percentage 13731 1727203862.96147: done checking for max_fail_percentage 13731 1727203862.96148: checking to see if all hosts have failed and the running result is not ok 13731 1727203862.96148: done checking to see if all hosts have failed 13731 1727203862.96149: getting the remaining hosts for this loop 13731 1727203862.96151: done getting the remaining hosts for this loop 13731 1727203862.96154: getting the next task for host managed-node3 13731 1727203862.96167: done getting next task for host managed-node3 13731 1727203862.96170: ^ task is: TASK: Reset bond options to assert 13731 1727203862.96172: ^ state is: HOST STATE: block=4, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203862.96184: getting variables 13731 1727203862.96186: in VariableManager get_vars() 13731 1727203862.96228: Calling all_inventory to load vars for managed-node3 13731 1727203862.96231: Calling groups_inventory to load vars for managed-node3 13731 1727203862.96233: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203862.96245: Calling all_plugins_play to load vars for managed-node3 13731 1727203862.96248: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203862.96251: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.03986: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.05566: done with get_vars() 13731 1727203863.05602: done getting variables 13731 1727203863.05662: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Reset bond options to assert] ******************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond_options.yml:59 Tuesday 24 September 2024 14:51:03 -0400 (0:00:00.532) 0:00:41.287 ***** 13731 1727203863.05689: entering _queue_task() for managed-node3/set_fact 13731 1727203863.06142: worker is 1 (out of 1 available) 13731 1727203863.06155: exiting _queue_task() for managed-node3/set_fact 13731 1727203863.06282: done queuing things up, now waiting for results queue to drain 13731 1727203863.06286: waiting for pending results... 13731 1727203863.06601: running TaskExecutor() for managed-node3/TASK: Reset bond options to assert 13731 1727203863.06630: in run() - task 028d2410-947f-82dc-c122-00000000000f 13731 1727203863.06696: variable 'ansible_search_path' from source: unknown 13731 1727203863.06710: calling self._execute() 13731 1727203863.06829: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.06842: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.06856: variable 'omit' from source: magic vars 13731 1727203863.07297: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.07350: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.07355: variable 'omit' from source: magic vars 13731 1727203863.07379: variable 'omit' from source: magic vars 13731 1727203863.07421: variable 'dhcp_interface1' from source: play vars 13731 1727203863.07506: variable 'dhcp_interface1' from source: play vars 13731 1727203863.07570: variable 'omit' from source: magic vars 13731 1727203863.07592: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203863.07630: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.07656: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203863.07691: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.07709: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.07785: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.07788: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.07791: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.07882: Set connection var ansible_pipelining to False 13731 1727203863.07980: Set connection var ansible_shell_type to sh 13731 1727203863.07984: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.07987: Set connection var ansible_connection to ssh 13731 1727203863.07989: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.07991: Set connection var ansible_timeout to 10 13731 1727203863.08084: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.08088: variable 'ansible_connection' from source: unknown 13731 1727203863.08091: variable 'ansible_module_compression' from source: unknown 13731 1727203863.08093: variable 'ansible_shell_type' from source: unknown 13731 1727203863.08096: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.08098: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.08100: variable 'ansible_pipelining' from source: unknown 13731 1727203863.08103: variable 'ansible_timeout' from source: unknown 13731 1727203863.08105: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.08223: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.08227: variable 'omit' from source: magic vars 13731 1727203863.08230: starting attempt loop 13731 1727203863.08232: running the handler 13731 1727203863.08234: handler run complete 13731 1727203863.08245: attempt loop complete, returning result 13731 1727203863.08253: _execute() done 13731 1727203863.08281: dumping result to json 13731 1727203863.08285: done dumping result, returning 13731 1727203863.08287: done running TaskExecutor() for managed-node3/TASK: Reset bond options to assert [028d2410-947f-82dc-c122-00000000000f] 13731 1727203863.08294: sending task result for task 028d2410-947f-82dc-c122-00000000000f ok: [managed-node3] => { "ansible_facts": { "bond_options_to_assert": [ { "key": "mode", "value": "active-backup" }, { "key": "arp_interval", "value": "60" }, { "key": "arp_ip_target", "value": "192.0.2.128" }, { "key": "arp_validate", "value": "none" }, { "key": "primary", "value": "test1" } ] }, "changed": false } 13731 1727203863.08619: no more pending results, returning what we have 13731 1727203863.08623: results queue empty 13731 1727203863.08623: checking for any_errors_fatal 13731 1727203863.08635: done checking for any_errors_fatal 13731 1727203863.08636: checking for max_fail_percentage 13731 1727203863.08638: done checking for max_fail_percentage 13731 1727203863.08639: checking to see if all hosts have failed and the running result is not ok 13731 1727203863.08640: done checking to see if all hosts have failed 13731 1727203863.08640: getting the remaining hosts for this loop 13731 1727203863.08643: done getting the remaining hosts for this loop 13731 1727203863.08646: getting the next task for host managed-node3 13731 1727203863.08656: done getting next task for host managed-node3 13731 1727203863.08661: ^ task is: TASK: Include the task 'run_test.yml' 13731 1727203863.08663: ^ state is: HOST STATE: block=5, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203863.08666: getting variables 13731 1727203863.08668: in VariableManager get_vars() 13731 1727203863.08709: Calling all_inventory to load vars for managed-node3 13731 1727203863.08712: Calling groups_inventory to load vars for managed-node3 13731 1727203863.08715: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.08726: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.08729: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.08731: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.09327: done sending task result for task 028d2410-947f-82dc-c122-00000000000f 13731 1727203863.09331: WORKER PROCESS EXITING 13731 1727203863.10322: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.11986: done with get_vars() 13731 1727203863.12023: done getting variables TASK [Include the task 'run_test.yml'] ***************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond_options.yml:72 Tuesday 24 September 2024 14:51:03 -0400 (0:00:00.064) 0:00:41.352 ***** 13731 1727203863.12130: entering _queue_task() for managed-node3/include_tasks 13731 1727203863.12519: worker is 1 (out of 1 available) 13731 1727203863.12532: exiting _queue_task() for managed-node3/include_tasks 13731 1727203863.12545: done queuing things up, now waiting for results queue to drain 13731 1727203863.12552: waiting for pending results... 13731 1727203863.12805: running TaskExecutor() for managed-node3/TASK: Include the task 'run_test.yml' 13731 1727203863.12929: in run() - task 028d2410-947f-82dc-c122-000000000011 13731 1727203863.12951: variable 'ansible_search_path' from source: unknown 13731 1727203863.13002: calling self._execute() 13731 1727203863.13123: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.13136: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.13148: variable 'omit' from source: magic vars 13731 1727203863.13565: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.13664: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.13668: _execute() done 13731 1727203863.13671: dumping result to json 13731 1727203863.13673: done dumping result, returning 13731 1727203863.13676: done running TaskExecutor() for managed-node3/TASK: Include the task 'run_test.yml' [028d2410-947f-82dc-c122-000000000011] 13731 1727203863.13679: sending task result for task 028d2410-947f-82dc-c122-000000000011 13731 1727203863.14008: no more pending results, returning what we have 13731 1727203863.14013: in VariableManager get_vars() 13731 1727203863.14050: Calling all_inventory to load vars for managed-node3 13731 1727203863.14053: Calling groups_inventory to load vars for managed-node3 13731 1727203863.14055: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.14068: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.14071: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.14077: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.14595: done sending task result for task 028d2410-947f-82dc-c122-000000000011 13731 1727203863.14599: WORKER PROCESS EXITING 13731 1727203863.15528: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.17102: done with get_vars() 13731 1727203863.17123: variable 'ansible_search_path' from source: unknown 13731 1727203863.17143: we have included files to process 13731 1727203863.17145: generating all_blocks data 13731 1727203863.17149: done generating all_blocks data 13731 1727203863.17155: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml 13731 1727203863.17156: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml 13731 1727203863.17161: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml 13731 1727203863.17626: in VariableManager get_vars() 13731 1727203863.17648: done with get_vars() 13731 1727203863.17699: in VariableManager get_vars() 13731 1727203863.17722: done with get_vars() 13731 1727203863.17768: in VariableManager get_vars() 13731 1727203863.17795: done with get_vars() 13731 1727203863.17838: in VariableManager get_vars() 13731 1727203863.17858: done with get_vars() 13731 1727203863.17909: in VariableManager get_vars() 13731 1727203863.17929: done with get_vars() 13731 1727203863.18389: in VariableManager get_vars() 13731 1727203863.18408: done with get_vars() 13731 1727203863.18431: done processing included file 13731 1727203863.18434: iterating over new_blocks loaded from include file 13731 1727203863.18436: in VariableManager get_vars() 13731 1727203863.18461: done with get_vars() 13731 1727203863.18463: filtering new block on tags 13731 1727203863.18557: done filtering new block on tags 13731 1727203863.18560: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml for managed-node3 13731 1727203863.18566: extending task lists for all hosts with included blocks 13731 1727203863.18605: done extending task lists 13731 1727203863.18609: done processing included files 13731 1727203863.18609: results queue empty 13731 1727203863.18610: checking for any_errors_fatal 13731 1727203863.18613: done checking for any_errors_fatal 13731 1727203863.18614: checking for max_fail_percentage 13731 1727203863.18615: done checking for max_fail_percentage 13731 1727203863.18616: checking to see if all hosts have failed and the running result is not ok 13731 1727203863.18617: done checking to see if all hosts have failed 13731 1727203863.18617: getting the remaining hosts for this loop 13731 1727203863.18618: done getting the remaining hosts for this loop 13731 1727203863.18621: getting the next task for host managed-node3 13731 1727203863.18624: done getting next task for host managed-node3 13731 1727203863.18626: ^ task is: TASK: TEST: {{ lsr_description }} 13731 1727203863.18629: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203863.18631: getting variables 13731 1727203863.18632: in VariableManager get_vars() 13731 1727203863.18643: Calling all_inventory to load vars for managed-node3 13731 1727203863.18646: Calling groups_inventory to load vars for managed-node3 13731 1727203863.18648: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.18655: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.18657: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.18660: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.20071: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.22822: done with get_vars() 13731 1727203863.22854: done getting variables 13731 1727203863.22924: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203863.23206: variable 'lsr_description' from source: include params TASK [TEST: Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:5 Tuesday 24 September 2024 14:51:03 -0400 (0:00:00.112) 0:00:41.464 ***** 13731 1727203863.23354: entering _queue_task() for managed-node3/debug 13731 1727203863.23910: worker is 1 (out of 1 available) 13731 1727203863.23921: exiting _queue_task() for managed-node3/debug 13731 1727203863.23935: done queuing things up, now waiting for results queue to drain 13731 1727203863.23937: waiting for pending results... 13731 1727203863.24451: running TaskExecutor() for managed-node3/TASK: TEST: Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device. 13731 1727203863.24508: in run() - task 028d2410-947f-82dc-c122-0000000008ea 13731 1727203863.24529: variable 'ansible_search_path' from source: unknown 13731 1727203863.24538: variable 'ansible_search_path' from source: unknown 13731 1727203863.24589: calling self._execute() 13731 1727203863.24706: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.24718: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.24733: variable 'omit' from source: magic vars 13731 1727203863.25140: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.25158: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.25199: variable 'omit' from source: magic vars 13731 1727203863.25215: variable 'omit' from source: magic vars 13731 1727203863.25327: variable 'lsr_description' from source: include params 13731 1727203863.25355: variable 'omit' from source: magic vars 13731 1727203863.25417: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203863.25453: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.25525: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203863.25530: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.25532: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.25558: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.25567: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.25577: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.25693: Set connection var ansible_pipelining to False 13731 1727203863.25704: Set connection var ansible_shell_type to sh 13731 1727203863.25714: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.25742: Set connection var ansible_connection to ssh 13731 1727203863.25746: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.25747: Set connection var ansible_timeout to 10 13731 1727203863.25770: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.25779: variable 'ansible_connection' from source: unknown 13731 1727203863.25852: variable 'ansible_module_compression' from source: unknown 13731 1727203863.25855: variable 'ansible_shell_type' from source: unknown 13731 1727203863.25857: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.25859: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.25861: variable 'ansible_pipelining' from source: unknown 13731 1727203863.25863: variable 'ansible_timeout' from source: unknown 13731 1727203863.25865: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.25961: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.25979: variable 'omit' from source: magic vars 13731 1727203863.25990: starting attempt loop 13731 1727203863.25997: running the handler 13731 1727203863.26070: handler run complete 13731 1727203863.26079: attempt loop complete, returning result 13731 1727203863.26177: _execute() done 13731 1727203863.26183: dumping result to json 13731 1727203863.26186: done dumping result, returning 13731 1727203863.26188: done running TaskExecutor() for managed-node3/TASK: TEST: Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device. [028d2410-947f-82dc-c122-0000000008ea] 13731 1727203863.26191: sending task result for task 028d2410-947f-82dc-c122-0000000008ea 13731 1727203863.26267: done sending task result for task 028d2410-947f-82dc-c122-0000000008ea 13731 1727203863.26271: WORKER PROCESS EXITING ok: [managed-node3] => {} MSG: ########## Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device. ########## 13731 1727203863.26340: no more pending results, returning what we have 13731 1727203863.26347: results queue empty 13731 1727203863.26348: checking for any_errors_fatal 13731 1727203863.26350: done checking for any_errors_fatal 13731 1727203863.26351: checking for max_fail_percentage 13731 1727203863.26353: done checking for max_fail_percentage 13731 1727203863.26354: checking to see if all hosts have failed and the running result is not ok 13731 1727203863.26354: done checking to see if all hosts have failed 13731 1727203863.26355: getting the remaining hosts for this loop 13731 1727203863.26357: done getting the remaining hosts for this loop 13731 1727203863.26362: getting the next task for host managed-node3 13731 1727203863.26373: done getting next task for host managed-node3 13731 1727203863.26378: ^ task is: TASK: Show item 13731 1727203863.26382: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203863.26387: getting variables 13731 1727203863.26389: in VariableManager get_vars() 13731 1727203863.26620: Calling all_inventory to load vars for managed-node3 13731 1727203863.26623: Calling groups_inventory to load vars for managed-node3 13731 1727203863.26625: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.26639: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.26642: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.26645: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.28227: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.29899: done with get_vars() 13731 1727203863.29932: done getting variables 13731 1727203863.29999: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Show item] *************************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:9 Tuesday 24 September 2024 14:51:03 -0400 (0:00:00.066) 0:00:41.531 ***** 13731 1727203863.30043: entering _queue_task() for managed-node3/debug 13731 1727203863.30418: worker is 1 (out of 1 available) 13731 1727203863.30431: exiting _queue_task() for managed-node3/debug 13731 1727203863.30442: done queuing things up, now waiting for results queue to drain 13731 1727203863.30444: waiting for pending results... 13731 1727203863.30893: running TaskExecutor() for managed-node3/TASK: Show item 13731 1727203863.30898: in run() - task 028d2410-947f-82dc-c122-0000000008eb 13731 1727203863.30908: variable 'ansible_search_path' from source: unknown 13731 1727203863.30918: variable 'ansible_search_path' from source: unknown 13731 1727203863.30980: variable 'omit' from source: magic vars 13731 1727203863.31153: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.31172: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.31190: variable 'omit' from source: magic vars 13731 1727203863.31567: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.31678: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.31681: variable 'omit' from source: magic vars 13731 1727203863.31683: variable 'omit' from source: magic vars 13731 1727203863.31686: variable 'item' from source: unknown 13731 1727203863.31747: variable 'item' from source: unknown 13731 1727203863.31774: variable 'omit' from source: magic vars 13731 1727203863.31893: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203863.31897: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.31899: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203863.31914: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.31930: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.31968: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.31979: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.31989: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.32111: Set connection var ansible_pipelining to False 13731 1727203863.32123: Set connection var ansible_shell_type to sh 13731 1727203863.32181: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.32184: Set connection var ansible_connection to ssh 13731 1727203863.32186: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.32188: Set connection var ansible_timeout to 10 13731 1727203863.32190: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.32191: variable 'ansible_connection' from source: unknown 13731 1727203863.32193: variable 'ansible_module_compression' from source: unknown 13731 1727203863.32196: variable 'ansible_shell_type' from source: unknown 13731 1727203863.32203: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.32208: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.32220: variable 'ansible_pipelining' from source: unknown 13731 1727203863.32226: variable 'ansible_timeout' from source: unknown 13731 1727203863.32232: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.32386: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.32406: variable 'omit' from source: magic vars 13731 1727203863.32416: starting attempt loop 13731 1727203863.32437: running the handler 13731 1727203863.32490: variable 'lsr_description' from source: include params 13731 1727203863.32657: variable 'lsr_description' from source: include params 13731 1727203863.32663: handler run complete 13731 1727203863.32665: attempt loop complete, returning result 13731 1727203863.32667: variable 'item' from source: unknown 13731 1727203863.32710: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_description) => { "ansible_loop_var": "item", "item": "lsr_description", "lsr_description": "Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device." } 13731 1727203863.33282: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.33285: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.33288: variable 'omit' from source: magic vars 13731 1727203863.33481: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.33484: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.33487: variable 'omit' from source: magic vars 13731 1727203863.33500: variable 'omit' from source: magic vars 13731 1727203863.33552: variable 'item' from source: unknown 13731 1727203863.33635: variable 'item' from source: unknown 13731 1727203863.33682: variable 'omit' from source: magic vars 13731 1727203863.33692: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.33712: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.33725: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.33809: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.33812: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.33815: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.33857: Set connection var ansible_pipelining to False 13731 1727203863.33888: Set connection var ansible_shell_type to sh 13731 1727203863.33900: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.33913: Set connection var ansible_connection to ssh 13731 1727203863.33927: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.33937: Set connection var ansible_timeout to 10 13731 1727203863.33970: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.33982: variable 'ansible_connection' from source: unknown 13731 1727203863.34028: variable 'ansible_module_compression' from source: unknown 13731 1727203863.34032: variable 'ansible_shell_type' from source: unknown 13731 1727203863.34034: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.34036: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.34038: variable 'ansible_pipelining' from source: unknown 13731 1727203863.34040: variable 'ansible_timeout' from source: unknown 13731 1727203863.34042: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.34144: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.34162: variable 'omit' from source: magic vars 13731 1727203863.34182: starting attempt loop 13731 1727203863.34187: running the handler 13731 1727203863.34247: variable 'lsr_setup' from source: include params 13731 1727203863.34310: variable 'lsr_setup' from source: include params 13731 1727203863.34366: handler run complete 13731 1727203863.34399: attempt loop complete, returning result 13731 1727203863.34468: variable 'item' from source: unknown 13731 1727203863.34507: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_setup) => { "ansible_loop_var": "item", "item": "lsr_setup", "lsr_setup": [ "tasks/create_test_interfaces_with_dhcp.yml", "tasks/assert_dhcp_device_present.yml" ] } 13731 1727203863.34698: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.34782: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.34786: variable 'omit' from source: magic vars 13731 1727203863.34909: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.34926: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.34936: variable 'omit' from source: magic vars 13731 1727203863.34955: variable 'omit' from source: magic vars 13731 1727203863.35002: variable 'item' from source: unknown 13731 1727203863.35080: variable 'item' from source: unknown 13731 1727203863.35102: variable 'omit' from source: magic vars 13731 1727203863.35128: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.35229: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.35233: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.35236: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.35238: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.35241: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.35283: Set connection var ansible_pipelining to False 13731 1727203863.35294: Set connection var ansible_shell_type to sh 13731 1727203863.35303: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.35313: Set connection var ansible_connection to ssh 13731 1727203863.35322: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.35334: Set connection var ansible_timeout to 10 13731 1727203863.35368: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.35380: variable 'ansible_connection' from source: unknown 13731 1727203863.35450: variable 'ansible_module_compression' from source: unknown 13731 1727203863.35455: variable 'ansible_shell_type' from source: unknown 13731 1727203863.35457: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.35464: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.35466: variable 'ansible_pipelining' from source: unknown 13731 1727203863.35469: variable 'ansible_timeout' from source: unknown 13731 1727203863.35471: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.35561: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.35565: variable 'omit' from source: magic vars 13731 1727203863.35567: starting attempt loop 13731 1727203863.35569: running the handler 13731 1727203863.35596: variable 'lsr_test' from source: include params 13731 1727203863.35667: variable 'lsr_test' from source: include params 13731 1727203863.35693: handler run complete 13731 1727203863.35780: attempt loop complete, returning result 13731 1727203863.35783: variable 'item' from source: unknown 13731 1727203863.35807: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_test) => { "ansible_loop_var": "item", "item": "lsr_test", "lsr_test": [ "tasks/create_bond_profile_reconfigure.yml" ] } 13731 1727203863.36050: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.36053: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.36056: variable 'omit' from source: magic vars 13731 1727203863.36285: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.36288: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.36290: variable 'omit' from source: magic vars 13731 1727203863.36292: variable 'omit' from source: magic vars 13731 1727203863.36381: variable 'item' from source: unknown 13731 1727203863.36610: variable 'item' from source: unknown 13731 1727203863.36613: variable 'omit' from source: magic vars 13731 1727203863.36616: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.36618: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.36620: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.36622: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.36624: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.36626: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.36715: Set connection var ansible_pipelining to False 13731 1727203863.36726: Set connection var ansible_shell_type to sh 13731 1727203863.36734: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.36741: Set connection var ansible_connection to ssh 13731 1727203863.36748: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.36755: Set connection var ansible_timeout to 10 13731 1727203863.36778: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.36786: variable 'ansible_connection' from source: unknown 13731 1727203863.36828: variable 'ansible_module_compression' from source: unknown 13731 1727203863.36831: variable 'ansible_shell_type' from source: unknown 13731 1727203863.36834: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.36836: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.36838: variable 'ansible_pipelining' from source: unknown 13731 1727203863.36839: variable 'ansible_timeout' from source: unknown 13731 1727203863.36841: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.36927: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.36950: variable 'omit' from source: magic vars 13731 1727203863.37046: starting attempt loop 13731 1727203863.37049: running the handler 13731 1727203863.37052: variable 'lsr_assert' from source: include params 13731 1727203863.37064: variable 'lsr_assert' from source: include params 13731 1727203863.37087: handler run complete 13731 1727203863.37106: attempt loop complete, returning result 13731 1727203863.37127: variable 'item' from source: unknown 13731 1727203863.37205: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_assert) => { "ansible_loop_var": "item", "item": "lsr_assert", "lsr_assert": [ "tasks/assert_bond_options.yml" ] } 13731 1727203863.37373: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.37379: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.37482: variable 'omit' from source: magic vars 13731 1727203863.37608: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.37617: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.37624: variable 'omit' from source: magic vars 13731 1727203863.37638: variable 'omit' from source: magic vars 13731 1727203863.37679: variable 'item' from source: unknown 13731 1727203863.37755: variable 'item' from source: unknown 13731 1727203863.37782: variable 'omit' from source: magic vars 13731 1727203863.37819: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.37821: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.37828: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.37915: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.37918: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.37920: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.37929: Set connection var ansible_pipelining to False 13731 1727203863.37941: Set connection var ansible_shell_type to sh 13731 1727203863.37952: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.37966: Set connection var ansible_connection to ssh 13731 1727203863.37980: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.37992: Set connection var ansible_timeout to 10 13731 1727203863.38015: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.38029: variable 'ansible_connection' from source: unknown 13731 1727203863.38040: variable 'ansible_module_compression' from source: unknown 13731 1727203863.38046: variable 'ansible_shell_type' from source: unknown 13731 1727203863.38052: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.38058: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.38068: variable 'ansible_pipelining' from source: unknown 13731 1727203863.38073: variable 'ansible_timeout' from source: unknown 13731 1727203863.38082: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.38241: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.38244: variable 'omit' from source: magic vars 13731 1727203863.38247: starting attempt loop 13731 1727203863.38250: running the handler 13731 1727203863.38301: handler run complete 13731 1727203863.38316: attempt loop complete, returning result 13731 1727203863.38331: variable 'item' from source: unknown 13731 1727203863.38457: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_assert_when) => { "ansible_loop_var": "item", "item": "lsr_assert_when", "lsr_assert_when": "VARIABLE IS NOT DEFINED!: 'lsr_assert_when' is undefined" } 13731 1727203863.38690: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.38693: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.38695: variable 'omit' from source: magic vars 13731 1727203863.38732: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.38741: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.38748: variable 'omit' from source: magic vars 13731 1727203863.38766: variable 'omit' from source: magic vars 13731 1727203863.38816: variable 'item' from source: unknown 13731 1727203863.38884: variable 'item' from source: unknown 13731 1727203863.38906: variable 'omit' from source: magic vars 13731 1727203863.38929: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.38939: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.38947: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.38962: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.38968: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.38973: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.39045: Set connection var ansible_pipelining to False 13731 1727203863.39054: Set connection var ansible_shell_type to sh 13731 1727203863.39064: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.39073: Set connection var ansible_connection to ssh 13731 1727203863.39085: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.39123: Set connection var ansible_timeout to 10 13731 1727203863.39129: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.39134: variable 'ansible_connection' from source: unknown 13731 1727203863.39140: variable 'ansible_module_compression' from source: unknown 13731 1727203863.39148: variable 'ansible_shell_type' from source: unknown 13731 1727203863.39155: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.39231: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.39234: variable 'ansible_pipelining' from source: unknown 13731 1727203863.39236: variable 'ansible_timeout' from source: unknown 13731 1727203863.39239: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.39290: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.39300: variable 'omit' from source: magic vars 13731 1727203863.39339: starting attempt loop 13731 1727203863.39342: running the handler 13731 1727203863.39344: variable 'lsr_fail_debug' from source: play vars 13731 1727203863.39401: variable 'lsr_fail_debug' from source: play vars 13731 1727203863.39420: handler run complete 13731 1727203863.39435: attempt loop complete, returning result 13731 1727203863.39465: variable 'item' from source: unknown 13731 1727203863.39556: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_fail_debug) => { "ansible_loop_var": "item", "item": "lsr_fail_debug", "lsr_fail_debug": [ "__network_connections_result" ] } 13731 1727203863.39881: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.39884: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.39886: variable 'omit' from source: magic vars 13731 1727203863.39888: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.39890: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.39892: variable 'omit' from source: magic vars 13731 1727203863.39894: variable 'omit' from source: magic vars 13731 1727203863.39936: variable 'item' from source: unknown 13731 1727203863.40011: variable 'item' from source: unknown 13731 1727203863.40033: variable 'omit' from source: magic vars 13731 1727203863.40057: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.40078: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.40094: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.40109: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.40121: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.40127: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.40199: Set connection var ansible_pipelining to False 13731 1727203863.40210: Set connection var ansible_shell_type to sh 13731 1727203863.40224: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.40238: Set connection var ansible_connection to ssh 13731 1727203863.40248: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.40339: Set connection var ansible_timeout to 10 13731 1727203863.40342: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.40344: variable 'ansible_connection' from source: unknown 13731 1727203863.40346: variable 'ansible_module_compression' from source: unknown 13731 1727203863.40348: variable 'ansible_shell_type' from source: unknown 13731 1727203863.40350: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.40352: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.40354: variable 'ansible_pipelining' from source: unknown 13731 1727203863.40356: variable 'ansible_timeout' from source: unknown 13731 1727203863.40358: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.40428: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.40445: variable 'omit' from source: magic vars 13731 1727203863.40453: starting attempt loop 13731 1727203863.40458: running the handler 13731 1727203863.40482: variable 'lsr_cleanup' from source: include params 13731 1727203863.40542: variable 'lsr_cleanup' from source: include params 13731 1727203863.40578: handler run complete 13731 1727203863.40598: attempt loop complete, returning result 13731 1727203863.40822: variable 'item' from source: unknown 13731 1727203863.40825: variable 'item' from source: unknown ok: [managed-node3] => (item=lsr_cleanup) => { "ansible_loop_var": "item", "item": "lsr_cleanup", "lsr_cleanup": [ "tasks/cleanup_bond_profile+device.yml", "tasks/remove_test_interfaces_with_dhcp.yml", "tasks/check_network_dns.yml" ] } 13731 1727203863.40909: dumping result to json 13731 1727203863.40916: done dumping result, returning 13731 1727203863.40928: done running TaskExecutor() for managed-node3/TASK: Show item [028d2410-947f-82dc-c122-0000000008eb] 13731 1727203863.40989: sending task result for task 028d2410-947f-82dc-c122-0000000008eb 13731 1727203863.41146: no more pending results, returning what we have 13731 1727203863.41151: results queue empty 13731 1727203863.41152: checking for any_errors_fatal 13731 1727203863.41158: done checking for any_errors_fatal 13731 1727203863.41162: checking for max_fail_percentage 13731 1727203863.41164: done checking for max_fail_percentage 13731 1727203863.41165: checking to see if all hosts have failed and the running result is not ok 13731 1727203863.41165: done checking to see if all hosts have failed 13731 1727203863.41166: getting the remaining hosts for this loop 13731 1727203863.41168: done getting the remaining hosts for this loop 13731 1727203863.41171: getting the next task for host managed-node3 13731 1727203863.41180: done getting next task for host managed-node3 13731 1727203863.41183: ^ task is: TASK: Include the task 'show_interfaces.yml' 13731 1727203863.41185: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203863.41190: getting variables 13731 1727203863.41191: in VariableManager get_vars() 13731 1727203863.41346: Calling all_inventory to load vars for managed-node3 13731 1727203863.41350: Calling groups_inventory to load vars for managed-node3 13731 1727203863.41352: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.41368: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.41371: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.41374: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.41492: done sending task result for task 028d2410-947f-82dc-c122-0000000008eb 13731 1727203863.41496: WORKER PROCESS EXITING 13731 1727203863.43243: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.45080: done with get_vars() 13731 1727203863.45113: done getting variables TASK [Include the task 'show_interfaces.yml'] ********************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:21 Tuesday 24 September 2024 14:51:03 -0400 (0:00:00.151) 0:00:41.683 ***** 13731 1727203863.45222: entering _queue_task() for managed-node3/include_tasks 13731 1727203863.45811: worker is 1 (out of 1 available) 13731 1727203863.45823: exiting _queue_task() for managed-node3/include_tasks 13731 1727203863.45833: done queuing things up, now waiting for results queue to drain 13731 1727203863.45835: waiting for pending results... 13731 1727203863.45953: running TaskExecutor() for managed-node3/TASK: Include the task 'show_interfaces.yml' 13731 1727203863.46084: in run() - task 028d2410-947f-82dc-c122-0000000008ec 13731 1727203863.46106: variable 'ansible_search_path' from source: unknown 13731 1727203863.46170: variable 'ansible_search_path' from source: unknown 13731 1727203863.46177: calling self._execute() 13731 1727203863.46285: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.46301: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.46316: variable 'omit' from source: magic vars 13731 1727203863.46817: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.46820: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.46823: _execute() done 13731 1727203863.46826: dumping result to json 13731 1727203863.46827: done dumping result, returning 13731 1727203863.46831: done running TaskExecutor() for managed-node3/TASK: Include the task 'show_interfaces.yml' [028d2410-947f-82dc-c122-0000000008ec] 13731 1727203863.46833: sending task result for task 028d2410-947f-82dc-c122-0000000008ec 13731 1727203863.46903: done sending task result for task 028d2410-947f-82dc-c122-0000000008ec 13731 1727203863.46906: WORKER PROCESS EXITING 13731 1727203863.46937: no more pending results, returning what we have 13731 1727203863.46944: in VariableManager get_vars() 13731 1727203863.46994: Calling all_inventory to load vars for managed-node3 13731 1727203863.46997: Calling groups_inventory to load vars for managed-node3 13731 1727203863.47000: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.47013: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.47016: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.47018: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.48727: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.50377: done with get_vars() 13731 1727203863.50407: variable 'ansible_search_path' from source: unknown 13731 1727203863.50408: variable 'ansible_search_path' from source: unknown 13731 1727203863.50448: we have included files to process 13731 1727203863.50449: generating all_blocks data 13731 1727203863.50451: done generating all_blocks data 13731 1727203863.50456: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml 13731 1727203863.50457: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml 13731 1727203863.50462: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml 13731 1727203863.50579: in VariableManager get_vars() 13731 1727203863.50602: done with get_vars() 13731 1727203863.50735: done processing included file 13731 1727203863.50737: iterating over new_blocks loaded from include file 13731 1727203863.50738: in VariableManager get_vars() 13731 1727203863.50754: done with get_vars() 13731 1727203863.50755: filtering new block on tags 13731 1727203863.50793: done filtering new block on tags 13731 1727203863.50796: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml for managed-node3 13731 1727203863.50801: extending task lists for all hosts with included blocks 13731 1727203863.51291: done extending task lists 13731 1727203863.51293: done processing included files 13731 1727203863.51294: results queue empty 13731 1727203863.51294: checking for any_errors_fatal 13731 1727203863.51302: done checking for any_errors_fatal 13731 1727203863.51302: checking for max_fail_percentage 13731 1727203863.51304: done checking for max_fail_percentage 13731 1727203863.51304: checking to see if all hosts have failed and the running result is not ok 13731 1727203863.51305: done checking to see if all hosts have failed 13731 1727203863.51306: getting the remaining hosts for this loop 13731 1727203863.51307: done getting the remaining hosts for this loop 13731 1727203863.51309: getting the next task for host managed-node3 13731 1727203863.51313: done getting next task for host managed-node3 13731 1727203863.51316: ^ task is: TASK: Include the task 'get_current_interfaces.yml' 13731 1727203863.51318: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203863.51320: getting variables 13731 1727203863.51321: in VariableManager get_vars() 13731 1727203863.51333: Calling all_inventory to load vars for managed-node3 13731 1727203863.51335: Calling groups_inventory to load vars for managed-node3 13731 1727203863.51337: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.51342: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.51345: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.51347: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.52673: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.54238: done with get_vars() 13731 1727203863.54267: done getting variables TASK [Include the task 'get_current_interfaces.yml'] *************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml:3 Tuesday 24 September 2024 14:51:03 -0400 (0:00:00.091) 0:00:41.774 ***** 13731 1727203863.54347: entering _queue_task() for managed-node3/include_tasks 13731 1727203863.54709: worker is 1 (out of 1 available) 13731 1727203863.54723: exiting _queue_task() for managed-node3/include_tasks 13731 1727203863.54736: done queuing things up, now waiting for results queue to drain 13731 1727203863.54737: waiting for pending results... 13731 1727203863.55282: running TaskExecutor() for managed-node3/TASK: Include the task 'get_current_interfaces.yml' 13731 1727203863.55291: in run() - task 028d2410-947f-82dc-c122-000000000913 13731 1727203863.55296: variable 'ansible_search_path' from source: unknown 13731 1727203863.55299: variable 'ansible_search_path' from source: unknown 13731 1727203863.55303: calling self._execute() 13731 1727203863.55313: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.55321: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.55331: variable 'omit' from source: magic vars 13731 1727203863.55768: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.55782: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.55788: _execute() done 13731 1727203863.55791: dumping result to json 13731 1727203863.55794: done dumping result, returning 13731 1727203863.55801: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_current_interfaces.yml' [028d2410-947f-82dc-c122-000000000913] 13731 1727203863.55807: sending task result for task 028d2410-947f-82dc-c122-000000000913 13731 1727203863.55928: done sending task result for task 028d2410-947f-82dc-c122-000000000913 13731 1727203863.55932: WORKER PROCESS EXITING 13731 1727203863.55987: no more pending results, returning what we have 13731 1727203863.55994: in VariableManager get_vars() 13731 1727203863.56041: Calling all_inventory to load vars for managed-node3 13731 1727203863.56044: Calling groups_inventory to load vars for managed-node3 13731 1727203863.56047: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.56062: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.56065: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.56069: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.57579: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.59803: done with get_vars() 13731 1727203863.59832: variable 'ansible_search_path' from source: unknown 13731 1727203863.59834: variable 'ansible_search_path' from source: unknown 13731 1727203863.59879: we have included files to process 13731 1727203863.59880: generating all_blocks data 13731 1727203863.59882: done generating all_blocks data 13731 1727203863.59884: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml 13731 1727203863.59885: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml 13731 1727203863.59887: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml 13731 1727203863.60624: done processing included file 13731 1727203863.60626: iterating over new_blocks loaded from include file 13731 1727203863.60628: in VariableManager get_vars() 13731 1727203863.60649: done with get_vars() 13731 1727203863.60651: filtering new block on tags 13731 1727203863.60692: done filtering new block on tags 13731 1727203863.60695: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml for managed-node3 13731 1727203863.60701: extending task lists for all hosts with included blocks 13731 1727203863.61066: done extending task lists 13731 1727203863.61067: done processing included files 13731 1727203863.61068: results queue empty 13731 1727203863.61069: checking for any_errors_fatal 13731 1727203863.61073: done checking for any_errors_fatal 13731 1727203863.61073: checking for max_fail_percentage 13731 1727203863.61074: done checking for max_fail_percentage 13731 1727203863.61077: checking to see if all hosts have failed and the running result is not ok 13731 1727203863.61078: done checking to see if all hosts have failed 13731 1727203863.61079: getting the remaining hosts for this loop 13731 1727203863.61080: done getting the remaining hosts for this loop 13731 1727203863.61083: getting the next task for host managed-node3 13731 1727203863.61087: done getting next task for host managed-node3 13731 1727203863.61089: ^ task is: TASK: Gather current interface info 13731 1727203863.61092: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203863.61095: getting variables 13731 1727203863.61096: in VariableManager get_vars() 13731 1727203863.61110: Calling all_inventory to load vars for managed-node3 13731 1727203863.61113: Calling groups_inventory to load vars for managed-node3 13731 1727203863.61115: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203863.61121: Calling all_plugins_play to load vars for managed-node3 13731 1727203863.61123: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203863.61126: Calling groups_plugins_play to load vars for managed-node3 13731 1727203863.63186: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203863.65631: done with get_vars() 13731 1727203863.65664: done getting variables 13731 1727203863.65715: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Gather current interface info] ******************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml:3 Tuesday 24 September 2024 14:51:03 -0400 (0:00:00.113) 0:00:41.888 ***** 13731 1727203863.65749: entering _queue_task() for managed-node3/command 13731 1727203863.66228: worker is 1 (out of 1 available) 13731 1727203863.66239: exiting _queue_task() for managed-node3/command 13731 1727203863.66250: done queuing things up, now waiting for results queue to drain 13731 1727203863.66253: waiting for pending results... 13731 1727203863.66950: running TaskExecutor() for managed-node3/TASK: Gather current interface info 13731 1727203863.66956: in run() - task 028d2410-947f-82dc-c122-00000000094e 13731 1727203863.66962: variable 'ansible_search_path' from source: unknown 13731 1727203863.66965: variable 'ansible_search_path' from source: unknown 13731 1727203863.66967: calling self._execute() 13731 1727203863.66970: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.66972: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.66978: variable 'omit' from source: magic vars 13731 1727203863.67454: variable 'ansible_distribution_major_version' from source: facts 13731 1727203863.67461: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203863.67464: variable 'omit' from source: magic vars 13731 1727203863.67466: variable 'omit' from source: magic vars 13731 1727203863.67501: variable 'omit' from source: magic vars 13731 1727203863.67584: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203863.67588: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203863.67603: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203863.67622: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.67820: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203863.67824: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203863.67827: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.67830: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.67864: Set connection var ansible_pipelining to False 13731 1727203863.67868: Set connection var ansible_shell_type to sh 13731 1727203863.67971: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203863.67978: Set connection var ansible_connection to ssh 13731 1727203863.67985: Set connection var ansible_shell_executable to /bin/sh 13731 1727203863.67995: Set connection var ansible_timeout to 10 13731 1727203863.68014: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.68018: variable 'ansible_connection' from source: unknown 13731 1727203863.68021: variable 'ansible_module_compression' from source: unknown 13731 1727203863.68024: variable 'ansible_shell_type' from source: unknown 13731 1727203863.68027: variable 'ansible_shell_executable' from source: unknown 13731 1727203863.68030: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203863.68032: variable 'ansible_pipelining' from source: unknown 13731 1727203863.68035: variable 'ansible_timeout' from source: unknown 13731 1727203863.68037: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203863.68325: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203863.68328: variable 'omit' from source: magic vars 13731 1727203863.68331: starting attempt loop 13731 1727203863.68333: running the handler 13731 1727203863.68335: _low_level_execute_command(): starting 13731 1727203863.68433: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203863.69083: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203863.69097: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203863.69109: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203863.69124: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203863.69137: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203863.69171: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203863.69240: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203863.69254: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203863.69283: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203863.69391: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203863.71055: stdout chunk (state=3): >>>/root <<< 13731 1727203863.71150: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203863.71155: stdout chunk (state=3): >>><<< 13731 1727203863.71164: stderr chunk (state=3): >>><<< 13731 1727203863.71227: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203863.71247: _low_level_execute_command(): starting 13731 1727203863.71252: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734 `" && echo ansible-tmp-1727203863.7122705-16884-130867275531734="` echo /root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734 `" ) && sleep 0' 13731 1727203863.72385: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203863.72390: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203863.72401: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203863.72404: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203863.72407: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203863.72409: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203863.72527: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203863.72531: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203863.72694: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203863.74550: stdout chunk (state=3): >>>ansible-tmp-1727203863.7122705-16884-130867275531734=/root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734 <<< 13731 1727203863.74679: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203863.74692: stdout chunk (state=3): >>><<< 13731 1727203863.74702: stderr chunk (state=3): >>><<< 13731 1727203863.74908: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203863.7122705-16884-130867275531734=/root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203863.74912: variable 'ansible_module_compression' from source: unknown 13731 1727203863.74914: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203863.75094: variable 'ansible_facts' from source: unknown 13731 1727203863.75185: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/AnsiballZ_command.py 13731 1727203863.75616: Sending initial data 13731 1727203863.75619: Sent initial data (156 bytes) 13731 1727203863.76588: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203863.76658: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203863.76703: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203863.76720: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203863.76735: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203863.76801: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203863.78339: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203863.78398: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203863.78473: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpe4f1twdn /root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/AnsiballZ_command.py <<< 13731 1727203863.78479: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/AnsiballZ_command.py" <<< 13731 1727203863.78525: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpe4f1twdn" to remote "/root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/AnsiballZ_command.py" <<< 13731 1727203863.79937: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203863.79941: stdout chunk (state=3): >>><<< 13731 1727203863.79944: stderr chunk (state=3): >>><<< 13731 1727203863.80044: done transferring module to remote 13731 1727203863.80048: _low_level_execute_command(): starting 13731 1727203863.80052: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/ /root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/AnsiballZ_command.py && sleep 0' 13731 1727203863.80800: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203863.80831: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203863.80846: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203863.80869: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203863.80937: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203863.82728: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203863.82740: stdout chunk (state=3): >>><<< 13731 1727203863.82782: stderr chunk (state=3): >>><<< 13731 1727203863.82786: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203863.82795: _low_level_execute_command(): starting 13731 1727203863.82804: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/AnsiballZ_command.py && sleep 0' 13731 1727203863.83488: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203863.83507: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203863.83522: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203863.83551: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203863.83664: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203863.83691: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203863.83769: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203863.99189: stdout chunk (state=3): >>> {"changed": true, "stdout": "bonding_masters\neth0\nlo", "stderr": "", "rc": 0, "cmd": ["ls", "-1"], "start": "2024-09-24 14:51:03.988101", "end": "2024-09-24 14:51:03.991331", "delta": "0:00:00.003230", "msg": "", "invocation": {"module_args": {"chdir": "/sys/class/net", "_raw_params": "ls -1", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203864.00945: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203864.00949: stdout chunk (state=3): >>><<< 13731 1727203864.00951: stderr chunk (state=3): >>><<< 13731 1727203864.01136: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "bonding_masters\neth0\nlo", "stderr": "", "rc": 0, "cmd": ["ls", "-1"], "start": "2024-09-24 14:51:03.988101", "end": "2024-09-24 14:51:03.991331", "delta": "0:00:00.003230", "msg": "", "invocation": {"module_args": {"chdir": "/sys/class/net", "_raw_params": "ls -1", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203864.01140: done with _execute_module (ansible.legacy.command, {'chdir': '/sys/class/net', '_raw_params': 'ls -1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203864.01142: _low_level_execute_command(): starting 13731 1727203864.01144: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203863.7122705-16884-130867275531734/ > /dev/null 2>&1 && sleep 0' 13731 1727203864.02179: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203864.02229: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.02253: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203864.02290: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.02371: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203864.02480: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203864.02503: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203864.02578: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203864.04684: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203864.04688: stdout chunk (state=3): >>><<< 13731 1727203864.04690: stderr chunk (state=3): >>><<< 13731 1727203864.04692: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203864.04817: handler run complete 13731 1727203864.04821: Evaluated conditional (False): False 13731 1727203864.04824: attempt loop complete, returning result 13731 1727203864.04826: _execute() done 13731 1727203864.04828: dumping result to json 13731 1727203864.04830: done dumping result, returning 13731 1727203864.04832: done running TaskExecutor() for managed-node3/TASK: Gather current interface info [028d2410-947f-82dc-c122-00000000094e] 13731 1727203864.04833: sending task result for task 028d2410-947f-82dc-c122-00000000094e ok: [managed-node3] => { "changed": false, "cmd": [ "ls", "-1" ], "delta": "0:00:00.003230", "end": "2024-09-24 14:51:03.991331", "rc": 0, "start": "2024-09-24 14:51:03.988101" } STDOUT: bonding_masters eth0 lo 13731 1727203864.05309: no more pending results, returning what we have 13731 1727203864.05315: results queue empty 13731 1727203864.05316: checking for any_errors_fatal 13731 1727203864.05318: done checking for any_errors_fatal 13731 1727203864.05318: checking for max_fail_percentage 13731 1727203864.05320: done checking for max_fail_percentage 13731 1727203864.05321: checking to see if all hosts have failed and the running result is not ok 13731 1727203864.05322: done checking to see if all hosts have failed 13731 1727203864.05323: getting the remaining hosts for this loop 13731 1727203864.05325: done getting the remaining hosts for this loop 13731 1727203864.05328: getting the next task for host managed-node3 13731 1727203864.05337: done getting next task for host managed-node3 13731 1727203864.05340: ^ task is: TASK: Set current_interfaces 13731 1727203864.05345: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203864.05350: getting variables 13731 1727203864.05352: in VariableManager get_vars() 13731 1727203864.05878: Calling all_inventory to load vars for managed-node3 13731 1727203864.05883: Calling groups_inventory to load vars for managed-node3 13731 1727203864.05885: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203864.05898: Calling all_plugins_play to load vars for managed-node3 13731 1727203864.05901: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203864.05904: Calling groups_plugins_play to load vars for managed-node3 13731 1727203864.06686: done sending task result for task 028d2410-947f-82dc-c122-00000000094e 13731 1727203864.06690: WORKER PROCESS EXITING 13731 1727203864.10026: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203864.12430: done with get_vars() 13731 1727203864.12465: done getting variables 13731 1727203864.12528: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set current_interfaces] ************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml:9 Tuesday 24 September 2024 14:51:04 -0400 (0:00:00.468) 0:00:42.356 ***** 13731 1727203864.12569: entering _queue_task() for managed-node3/set_fact 13731 1727203864.12949: worker is 1 (out of 1 available) 13731 1727203864.12962: exiting _queue_task() for managed-node3/set_fact 13731 1727203864.13085: done queuing things up, now waiting for results queue to drain 13731 1727203864.13087: waiting for pending results... 13731 1727203864.13299: running TaskExecutor() for managed-node3/TASK: Set current_interfaces 13731 1727203864.13453: in run() - task 028d2410-947f-82dc-c122-00000000094f 13731 1727203864.13481: variable 'ansible_search_path' from source: unknown 13731 1727203864.13490: variable 'ansible_search_path' from source: unknown 13731 1727203864.13534: calling self._execute() 13731 1727203864.13633: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.13647: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.13661: variable 'omit' from source: magic vars 13731 1727203864.14042: variable 'ansible_distribution_major_version' from source: facts 13731 1727203864.14069: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203864.14086: variable 'omit' from source: magic vars 13731 1727203864.14147: variable 'omit' from source: magic vars 13731 1727203864.14274: variable '_current_interfaces' from source: set_fact 13731 1727203864.14350: variable 'omit' from source: magic vars 13731 1727203864.14406: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203864.14508: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203864.14511: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203864.14514: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203864.14516: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203864.14549: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203864.14557: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.14569: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.14686: Set connection var ansible_pipelining to False 13731 1727203864.14698: Set connection var ansible_shell_type to sh 13731 1727203864.14708: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203864.14726: Set connection var ansible_connection to ssh 13731 1727203864.14737: Set connection var ansible_shell_executable to /bin/sh 13731 1727203864.14747: Set connection var ansible_timeout to 10 13731 1727203864.14835: variable 'ansible_shell_executable' from source: unknown 13731 1727203864.14839: variable 'ansible_connection' from source: unknown 13731 1727203864.14841: variable 'ansible_module_compression' from source: unknown 13731 1727203864.14843: variable 'ansible_shell_type' from source: unknown 13731 1727203864.14846: variable 'ansible_shell_executable' from source: unknown 13731 1727203864.14848: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.14850: variable 'ansible_pipelining' from source: unknown 13731 1727203864.14852: variable 'ansible_timeout' from source: unknown 13731 1727203864.14854: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.14991: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203864.15009: variable 'omit' from source: magic vars 13731 1727203864.15019: starting attempt loop 13731 1727203864.15026: running the handler 13731 1727203864.15049: handler run complete 13731 1727203864.15163: attempt loop complete, returning result 13731 1727203864.15166: _execute() done 13731 1727203864.15168: dumping result to json 13731 1727203864.15170: done dumping result, returning 13731 1727203864.15172: done running TaskExecutor() for managed-node3/TASK: Set current_interfaces [028d2410-947f-82dc-c122-00000000094f] 13731 1727203864.15174: sending task result for task 028d2410-947f-82dc-c122-00000000094f 13731 1727203864.15244: done sending task result for task 028d2410-947f-82dc-c122-00000000094f 13731 1727203864.15247: WORKER PROCESS EXITING ok: [managed-node3] => { "ansible_facts": { "current_interfaces": [ "bonding_masters", "eth0", "lo" ] }, "changed": false } 13731 1727203864.15313: no more pending results, returning what we have 13731 1727203864.15318: results queue empty 13731 1727203864.15319: checking for any_errors_fatal 13731 1727203864.15330: done checking for any_errors_fatal 13731 1727203864.15331: checking for max_fail_percentage 13731 1727203864.15333: done checking for max_fail_percentage 13731 1727203864.15334: checking to see if all hosts have failed and the running result is not ok 13731 1727203864.15335: done checking to see if all hosts have failed 13731 1727203864.15336: getting the remaining hosts for this loop 13731 1727203864.15338: done getting the remaining hosts for this loop 13731 1727203864.15342: getting the next task for host managed-node3 13731 1727203864.15353: done getting next task for host managed-node3 13731 1727203864.15356: ^ task is: TASK: Show current_interfaces 13731 1727203864.15364: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203864.15369: getting variables 13731 1727203864.15371: in VariableManager get_vars() 13731 1727203864.15419: Calling all_inventory to load vars for managed-node3 13731 1727203864.15423: Calling groups_inventory to load vars for managed-node3 13731 1727203864.15425: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203864.15438: Calling all_plugins_play to load vars for managed-node3 13731 1727203864.15441: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203864.15444: Calling groups_plugins_play to load vars for managed-node3 13731 1727203864.17317: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203864.20718: done with get_vars() 13731 1727203864.20755: done getting variables 13731 1727203864.20930: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Show current_interfaces] ************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml:5 Tuesday 24 September 2024 14:51:04 -0400 (0:00:00.083) 0:00:42.440 ***** 13731 1727203864.20965: entering _queue_task() for managed-node3/debug 13731 1727203864.21917: worker is 1 (out of 1 available) 13731 1727203864.21930: exiting _queue_task() for managed-node3/debug 13731 1727203864.21943: done queuing things up, now waiting for results queue to drain 13731 1727203864.21945: waiting for pending results... 13731 1727203864.22412: running TaskExecutor() for managed-node3/TASK: Show current_interfaces 13731 1727203864.22536: in run() - task 028d2410-947f-82dc-c122-000000000914 13731 1727203864.22550: variable 'ansible_search_path' from source: unknown 13731 1727203864.22554: variable 'ansible_search_path' from source: unknown 13731 1727203864.22597: calling self._execute() 13731 1727203864.22701: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.22707: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.22718: variable 'omit' from source: magic vars 13731 1727203864.23119: variable 'ansible_distribution_major_version' from source: facts 13731 1727203864.23131: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203864.23137: variable 'omit' from source: magic vars 13731 1727203864.23193: variable 'omit' from source: magic vars 13731 1727203864.23298: variable 'current_interfaces' from source: set_fact 13731 1727203864.23377: variable 'omit' from source: magic vars 13731 1727203864.23381: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203864.23414: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203864.23435: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203864.23454: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203864.23467: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203864.23534: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203864.23538: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.23540: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.23767: Set connection var ansible_pipelining to False 13731 1727203864.23776: Set connection var ansible_shell_type to sh 13731 1727203864.23779: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203864.23781: Set connection var ansible_connection to ssh 13731 1727203864.23783: Set connection var ansible_shell_executable to /bin/sh 13731 1727203864.23785: Set connection var ansible_timeout to 10 13731 1727203864.23878: variable 'ansible_shell_executable' from source: unknown 13731 1727203864.23885: variable 'ansible_connection' from source: unknown 13731 1727203864.23887: variable 'ansible_module_compression' from source: unknown 13731 1727203864.23890: variable 'ansible_shell_type' from source: unknown 13731 1727203864.23892: variable 'ansible_shell_executable' from source: unknown 13731 1727203864.23894: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.23896: variable 'ansible_pipelining' from source: unknown 13731 1727203864.23898: variable 'ansible_timeout' from source: unknown 13731 1727203864.23900: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.24382: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203864.24385: variable 'omit' from source: magic vars 13731 1727203864.24387: starting attempt loop 13731 1727203864.24390: running the handler 13731 1727203864.24394: handler run complete 13731 1727203864.24441: attempt loop complete, returning result 13731 1727203864.24444: _execute() done 13731 1727203864.24447: dumping result to json 13731 1727203864.24449: done dumping result, returning 13731 1727203864.24462: done running TaskExecutor() for managed-node3/TASK: Show current_interfaces [028d2410-947f-82dc-c122-000000000914] 13731 1727203864.24465: sending task result for task 028d2410-947f-82dc-c122-000000000914 13731 1727203864.24557: done sending task result for task 028d2410-947f-82dc-c122-000000000914 13731 1727203864.24562: WORKER PROCESS EXITING ok: [managed-node3] => {} MSG: current_interfaces: ['bonding_masters', 'eth0', 'lo'] 13731 1727203864.24620: no more pending results, returning what we have 13731 1727203864.24624: results queue empty 13731 1727203864.24625: checking for any_errors_fatal 13731 1727203864.24632: done checking for any_errors_fatal 13731 1727203864.24633: checking for max_fail_percentage 13731 1727203864.24634: done checking for max_fail_percentage 13731 1727203864.24635: checking to see if all hosts have failed and the running result is not ok 13731 1727203864.24636: done checking to see if all hosts have failed 13731 1727203864.24636: getting the remaining hosts for this loop 13731 1727203864.24639: done getting the remaining hosts for this loop 13731 1727203864.24642: getting the next task for host managed-node3 13731 1727203864.24651: done getting next task for host managed-node3 13731 1727203864.24655: ^ task is: TASK: Setup 13731 1727203864.24658: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203864.24666: getting variables 13731 1727203864.24668: in VariableManager get_vars() 13731 1727203864.24706: Calling all_inventory to load vars for managed-node3 13731 1727203864.24709: Calling groups_inventory to load vars for managed-node3 13731 1727203864.24711: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203864.24722: Calling all_plugins_play to load vars for managed-node3 13731 1727203864.24724: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203864.24726: Calling groups_plugins_play to load vars for managed-node3 13731 1727203864.26628: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203864.28478: done with get_vars() 13731 1727203864.28518: done getting variables TASK [Setup] ******************************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:24 Tuesday 24 September 2024 14:51:04 -0400 (0:00:00.076) 0:00:42.517 ***** 13731 1727203864.28626: entering _queue_task() for managed-node3/include_tasks 13731 1727203864.29215: worker is 1 (out of 1 available) 13731 1727203864.29225: exiting _queue_task() for managed-node3/include_tasks 13731 1727203864.29235: done queuing things up, now waiting for results queue to drain 13731 1727203864.29237: waiting for pending results... 13731 1727203864.29898: running TaskExecutor() for managed-node3/TASK: Setup 13731 1727203864.29904: in run() - task 028d2410-947f-82dc-c122-0000000008ed 13731 1727203864.29907: variable 'ansible_search_path' from source: unknown 13731 1727203864.29911: variable 'ansible_search_path' from source: unknown 13731 1727203864.29918: variable 'lsr_setup' from source: include params 13731 1727203864.30264: variable 'lsr_setup' from source: include params 13731 1727203864.30481: variable 'omit' from source: magic vars 13731 1727203864.30981: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.30986: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.30989: variable 'omit' from source: magic vars 13731 1727203864.31072: variable 'ansible_distribution_major_version' from source: facts 13731 1727203864.31084: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203864.31095: variable 'item' from source: unknown 13731 1727203864.31164: variable 'item' from source: unknown 13731 1727203864.31201: variable 'item' from source: unknown 13731 1727203864.31264: variable 'item' from source: unknown 13731 1727203864.31566: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.31569: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.31572: variable 'omit' from source: magic vars 13731 1727203864.31574: variable 'ansible_distribution_major_version' from source: facts 13731 1727203864.31578: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203864.31580: variable 'item' from source: unknown 13731 1727203864.31783: variable 'item' from source: unknown 13731 1727203864.31785: variable 'item' from source: unknown 13731 1727203864.31787: variable 'item' from source: unknown 13731 1727203864.31834: dumping result to json 13731 1727203864.31837: done dumping result, returning 13731 1727203864.31839: done running TaskExecutor() for managed-node3/TASK: Setup [028d2410-947f-82dc-c122-0000000008ed] 13731 1727203864.31841: sending task result for task 028d2410-947f-82dc-c122-0000000008ed 13731 1727203864.31878: done sending task result for task 028d2410-947f-82dc-c122-0000000008ed 13731 1727203864.31881: WORKER PROCESS EXITING 13731 1727203864.31905: no more pending results, returning what we have 13731 1727203864.31911: in VariableManager get_vars() 13731 1727203864.31953: Calling all_inventory to load vars for managed-node3 13731 1727203864.31956: Calling groups_inventory to load vars for managed-node3 13731 1727203864.31961: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203864.31974: Calling all_plugins_play to load vars for managed-node3 13731 1727203864.31979: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203864.31982: Calling groups_plugins_play to load vars for managed-node3 13731 1727203864.33527: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203864.35720: done with get_vars() 13731 1727203864.35744: variable 'ansible_search_path' from source: unknown 13731 1727203864.35745: variable 'ansible_search_path' from source: unknown 13731 1727203864.35793: variable 'ansible_search_path' from source: unknown 13731 1727203864.35794: variable 'ansible_search_path' from source: unknown 13731 1727203864.35819: we have included files to process 13731 1727203864.35820: generating all_blocks data 13731 1727203864.35821: done generating all_blocks data 13731 1727203864.35826: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml 13731 1727203864.35827: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml 13731 1727203864.35829: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml 13731 1727203864.37126: done processing included file 13731 1727203864.37129: iterating over new_blocks loaded from include file 13731 1727203864.37130: in VariableManager get_vars() 13731 1727203864.37150: done with get_vars() 13731 1727203864.37153: filtering new block on tags 13731 1727203864.37210: done filtering new block on tags 13731 1727203864.37213: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml for managed-node3 => (item=tasks/create_test_interfaces_with_dhcp.yml) 13731 1727203864.37219: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_dhcp_device_present.yml 13731 1727203864.37220: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_dhcp_device_present.yml 13731 1727203864.37224: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_dhcp_device_present.yml 13731 1727203864.37317: in VariableManager get_vars() 13731 1727203864.37339: done with get_vars() 13731 1727203864.37346: variable 'item' from source: include params 13731 1727203864.37456: variable 'item' from source: include params 13731 1727203864.37493: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml statically imported: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml 13731 1727203864.37577: in VariableManager get_vars() 13731 1727203864.37600: done with get_vars() 13731 1727203864.37733: in VariableManager get_vars() 13731 1727203864.37753: done with get_vars() 13731 1727203864.37761: variable 'item' from source: include params 13731 1727203864.37827: variable 'item' from source: include params 13731 1727203864.37857: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml statically imported: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml 13731 1727203864.38017: in VariableManager get_vars() 13731 1727203864.38038: done with get_vars() 13731 1727203864.38141: done processing included file 13731 1727203864.38143: iterating over new_blocks loaded from include file 13731 1727203864.38145: in VariableManager get_vars() 13731 1727203864.38163: done with get_vars() 13731 1727203864.38165: filtering new block on tags 13731 1727203864.38241: done filtering new block on tags 13731 1727203864.38244: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_dhcp_device_present.yml for managed-node3 => (item=tasks/assert_dhcp_device_present.yml) 13731 1727203864.38249: extending task lists for all hosts with included blocks 13731 1727203864.38820: done extending task lists 13731 1727203864.38821: done processing included files 13731 1727203864.38822: results queue empty 13731 1727203864.38823: checking for any_errors_fatal 13731 1727203864.38826: done checking for any_errors_fatal 13731 1727203864.38827: checking for max_fail_percentage 13731 1727203864.38828: done checking for max_fail_percentage 13731 1727203864.38829: checking to see if all hosts have failed and the running result is not ok 13731 1727203864.38830: done checking to see if all hosts have failed 13731 1727203864.38836: getting the remaining hosts for this loop 13731 1727203864.38838: done getting the remaining hosts for this loop 13731 1727203864.38840: getting the next task for host managed-node3 13731 1727203864.38844: done getting next task for host managed-node3 13731 1727203864.38846: ^ task is: TASK: Install dnsmasq 13731 1727203864.38849: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203864.38851: getting variables 13731 1727203864.38851: in VariableManager get_vars() 13731 1727203864.38864: Calling all_inventory to load vars for managed-node3 13731 1727203864.38867: Calling groups_inventory to load vars for managed-node3 13731 1727203864.38868: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203864.38874: Calling all_plugins_play to load vars for managed-node3 13731 1727203864.38877: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203864.38880: Calling groups_plugins_play to load vars for managed-node3 13731 1727203864.40043: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203864.41621: done with get_vars() 13731 1727203864.41648: done getting variables 13731 1727203864.41700: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Install dnsmasq] ********************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:3 Tuesday 24 September 2024 14:51:04 -0400 (0:00:00.131) 0:00:42.648 ***** 13731 1727203864.41735: entering _queue_task() for managed-node3/package 13731 1727203864.42162: worker is 1 (out of 1 available) 13731 1727203864.42178: exiting _queue_task() for managed-node3/package 13731 1727203864.42192: done queuing things up, now waiting for results queue to drain 13731 1727203864.42193: waiting for pending results... 13731 1727203864.42506: running TaskExecutor() for managed-node3/TASK: Install dnsmasq 13731 1727203864.42670: in run() - task 028d2410-947f-82dc-c122-000000000974 13731 1727203864.42677: variable 'ansible_search_path' from source: unknown 13731 1727203864.42681: variable 'ansible_search_path' from source: unknown 13731 1727203864.42711: calling self._execute() 13731 1727203864.42941: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.42945: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.42948: variable 'omit' from source: magic vars 13731 1727203864.43273: variable 'ansible_distribution_major_version' from source: facts 13731 1727203864.43328: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203864.43340: variable 'omit' from source: magic vars 13731 1727203864.43400: variable 'omit' from source: magic vars 13731 1727203864.43637: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203864.47642: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203864.47647: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203864.47726: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203864.47820: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203864.47918: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203864.48083: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203864.48404: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203864.48407: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203864.48409: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203864.48464: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203864.48684: variable '__network_is_ostree' from source: set_fact 13731 1727203864.48977: variable 'omit' from source: magic vars 13731 1727203864.48981: variable 'omit' from source: magic vars 13731 1727203864.48983: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203864.48986: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203864.48988: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203864.49089: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203864.49108: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203864.49142: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203864.49173: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.49185: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.49418: Set connection var ansible_pipelining to False 13731 1727203864.49436: Set connection var ansible_shell_type to sh 13731 1727203864.49449: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203864.49466: Set connection var ansible_connection to ssh 13731 1727203864.49481: Set connection var ansible_shell_executable to /bin/sh 13731 1727203864.49493: Set connection var ansible_timeout to 10 13731 1727203864.49522: variable 'ansible_shell_executable' from source: unknown 13731 1727203864.49530: variable 'ansible_connection' from source: unknown 13731 1727203864.49543: variable 'ansible_module_compression' from source: unknown 13731 1727203864.49551: variable 'ansible_shell_type' from source: unknown 13731 1727203864.49561: variable 'ansible_shell_executable' from source: unknown 13731 1727203864.49570: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203864.49649: variable 'ansible_pipelining' from source: unknown 13731 1727203864.49652: variable 'ansible_timeout' from source: unknown 13731 1727203864.49654: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203864.49892: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203864.49895: variable 'omit' from source: magic vars 13731 1727203864.49897: starting attempt loop 13731 1727203864.49899: running the handler 13731 1727203864.49901: variable 'ansible_facts' from source: unknown 13731 1727203864.49903: variable 'ansible_facts' from source: unknown 13731 1727203864.50067: _low_level_execute_command(): starting 13731 1727203864.50091: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203864.51748: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.51794: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203864.51813: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203864.51856: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203864.52013: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203864.53702: stdout chunk (state=3): >>>/root <<< 13731 1727203864.54067: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203864.54134: stdout chunk (state=3): >>><<< 13731 1727203864.54138: stderr chunk (state=3): >>><<< 13731 1727203864.54158: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203864.54496: _low_level_execute_command(): starting 13731 1727203864.54500: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708 `" && echo ansible-tmp-1727203864.542864-16989-32451340615708="` echo /root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708 `" ) && sleep 0' 13731 1727203864.55972: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203864.55978: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.55981: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203864.55985: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.56209: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203864.56293: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203864.58104: stdout chunk (state=3): >>>ansible-tmp-1727203864.542864-16989-32451340615708=/root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708 <<< 13731 1727203864.58208: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203864.58251: stderr chunk (state=3): >>><<< 13731 1727203864.58255: stdout chunk (state=3): >>><<< 13731 1727203864.58280: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203864.542864-16989-32451340615708=/root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203864.58314: variable 'ansible_module_compression' from source: unknown 13731 1727203864.58373: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.dnf-ZIP_DEFLATED 13731 1727203864.58420: variable 'ansible_facts' from source: unknown 13731 1727203864.58581: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/AnsiballZ_dnf.py 13731 1727203864.58839: Sending initial data 13731 1727203864.58843: Sent initial data (150 bytes) 13731 1727203864.60137: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203864.60141: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.60227: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203864.60236: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203864.60297: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203864.60463: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203864.62012: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203864.62033: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203864.62099: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmplgednu6s /root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/AnsiballZ_dnf.py <<< 13731 1727203864.62103: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/AnsiballZ_dnf.py" <<< 13731 1727203864.62236: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmplgednu6s" to remote "/root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/AnsiballZ_dnf.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/AnsiballZ_dnf.py" <<< 13731 1727203864.64096: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203864.64152: stderr chunk (state=3): >>><<< 13731 1727203864.64385: stdout chunk (state=3): >>><<< 13731 1727203864.64388: done transferring module to remote 13731 1727203864.64390: _low_level_execute_command(): starting 13731 1727203864.64393: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/ /root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/AnsiballZ_dnf.py && sleep 0' 13731 1727203864.65588: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203864.65612: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203864.65794: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203864.65826: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.65867: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203864.65898: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203864.66215: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203864.66255: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203864.68058: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203864.68096: stderr chunk (state=3): >>><<< 13731 1727203864.68104: stdout chunk (state=3): >>><<< 13731 1727203864.68129: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203864.68137: _low_level_execute_command(): starting 13731 1727203864.68150: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/AnsiballZ_dnf.py && sleep 0' 13731 1727203864.69457: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203864.69460: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.69544: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203864.69716: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203865.10243: stdout chunk (state=3): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["dnsmasq"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} <<< 13731 1727203865.14287: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203865.14568: stderr chunk (state=3): >>><<< 13731 1727203865.14572: stdout chunk (state=3): >>><<< 13731 1727203865.14587: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["dnsmasq"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203865.14591: done with _execute_module (ansible.legacy.dnf, {'name': 'dnsmasq', 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203865.14599: _low_level_execute_command(): starting 13731 1727203865.14601: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203864.542864-16989-32451340615708/ > /dev/null 2>&1 && sleep 0' 13731 1727203865.15687: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203865.15692: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203865.15695: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203865.15895: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203865.15940: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203865.16051: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203865.17884: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203865.17925: stderr chunk (state=3): >>><<< 13731 1727203865.18026: stdout chunk (state=3): >>><<< 13731 1727203865.18039: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203865.18042: handler run complete 13731 1727203865.18359: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203865.18587: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203865.18631: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203865.18667: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203865.18712: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203865.18790: variable '__install_status' from source: set_fact 13731 1727203865.18821: Evaluated conditional (__install_status is success): True 13731 1727203865.18843: attempt loop complete, returning result 13731 1727203865.18850: _execute() done 13731 1727203865.18855: dumping result to json 13731 1727203865.18864: done dumping result, returning 13731 1727203865.18880: done running TaskExecutor() for managed-node3/TASK: Install dnsmasq [028d2410-947f-82dc-c122-000000000974] 13731 1727203865.18890: sending task result for task 028d2410-947f-82dc-c122-000000000974 ok: [managed-node3] => { "attempts": 1, "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 13731 1727203865.19122: no more pending results, returning what we have 13731 1727203865.19128: results queue empty 13731 1727203865.19129: checking for any_errors_fatal 13731 1727203865.19130: done checking for any_errors_fatal 13731 1727203865.19131: checking for max_fail_percentage 13731 1727203865.19133: done checking for max_fail_percentage 13731 1727203865.19133: checking to see if all hosts have failed and the running result is not ok 13731 1727203865.19134: done checking to see if all hosts have failed 13731 1727203865.19135: getting the remaining hosts for this loop 13731 1727203865.19137: done getting the remaining hosts for this loop 13731 1727203865.19140: getting the next task for host managed-node3 13731 1727203865.19146: done getting next task for host managed-node3 13731 1727203865.19149: ^ task is: TASK: Install pgrep, sysctl 13731 1727203865.19152: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203865.19155: getting variables 13731 1727203865.19157: in VariableManager get_vars() 13731 1727203865.19197: Calling all_inventory to load vars for managed-node3 13731 1727203865.19200: Calling groups_inventory to load vars for managed-node3 13731 1727203865.19202: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203865.19213: Calling all_plugins_play to load vars for managed-node3 13731 1727203865.19216: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203865.19218: Calling groups_plugins_play to load vars for managed-node3 13731 1727203865.19923: done sending task result for task 028d2410-947f-82dc-c122-000000000974 13731 1727203865.19927: WORKER PROCESS EXITING 13731 1727203865.21590: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203865.24248: done with get_vars() 13731 1727203865.24287: done getting variables 13731 1727203865.24352: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Install pgrep, sysctl] *************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:17 Tuesday 24 September 2024 14:51:05 -0400 (0:00:00.826) 0:00:43.474 ***** 13731 1727203865.24389: entering _queue_task() for managed-node3/package 13731 1727203865.24806: worker is 1 (out of 1 available) 13731 1727203865.24817: exiting _queue_task() for managed-node3/package 13731 1727203865.24831: done queuing things up, now waiting for results queue to drain 13731 1727203865.24833: waiting for pending results... 13731 1727203865.25204: running TaskExecutor() for managed-node3/TASK: Install pgrep, sysctl 13731 1727203865.25479: in run() - task 028d2410-947f-82dc-c122-000000000975 13731 1727203865.25503: variable 'ansible_search_path' from source: unknown 13731 1727203865.25512: variable 'ansible_search_path' from source: unknown 13731 1727203865.25555: calling self._execute() 13731 1727203865.25727: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203865.25755: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203865.25759: variable 'omit' from source: magic vars 13731 1727203865.26159: variable 'ansible_distribution_major_version' from source: facts 13731 1727203865.26223: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203865.26417: variable 'ansible_os_family' from source: facts 13731 1727203865.26421: Evaluated conditional (ansible_os_family == 'RedHat'): True 13731 1727203865.26589: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203865.26900: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203865.26954: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203865.26997: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203865.27039: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203865.27128: variable 'ansible_distribution_major_version' from source: facts 13731 1727203865.27160: Evaluated conditional (ansible_distribution_major_version is version('6', '<=')): False 13731 1727203865.27169: when evaluation is False, skipping this task 13731 1727203865.27179: _execute() done 13731 1727203865.27186: dumping result to json 13731 1727203865.27199: done dumping result, returning 13731 1727203865.27210: done running TaskExecutor() for managed-node3/TASK: Install pgrep, sysctl [028d2410-947f-82dc-c122-000000000975] 13731 1727203865.27219: sending task result for task 028d2410-947f-82dc-c122-000000000975 skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version is version('6', '<=')", "skip_reason": "Conditional result was False" } 13731 1727203865.27374: no more pending results, returning what we have 13731 1727203865.27381: results queue empty 13731 1727203865.27382: checking for any_errors_fatal 13731 1727203865.27395: done checking for any_errors_fatal 13731 1727203865.27396: checking for max_fail_percentage 13731 1727203865.27398: done checking for max_fail_percentage 13731 1727203865.27399: checking to see if all hosts have failed and the running result is not ok 13731 1727203865.27399: done checking to see if all hosts have failed 13731 1727203865.27400: getting the remaining hosts for this loop 13731 1727203865.27403: done getting the remaining hosts for this loop 13731 1727203865.27407: getting the next task for host managed-node3 13731 1727203865.27416: done getting next task for host managed-node3 13731 1727203865.27419: ^ task is: TASK: Install pgrep, sysctl 13731 1727203865.27423: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203865.27427: getting variables 13731 1727203865.27429: in VariableManager get_vars() 13731 1727203865.27470: Calling all_inventory to load vars for managed-node3 13731 1727203865.27474: Calling groups_inventory to load vars for managed-node3 13731 1727203865.27780: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203865.27791: Calling all_plugins_play to load vars for managed-node3 13731 1727203865.27795: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203865.27798: Calling groups_plugins_play to load vars for managed-node3 13731 1727203865.28491: done sending task result for task 028d2410-947f-82dc-c122-000000000975 13731 1727203865.28495: WORKER PROCESS EXITING 13731 1727203865.29858: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203865.32295: done with get_vars() 13731 1727203865.32330: done getting variables 13731 1727203865.32395: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Install pgrep, sysctl] *************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:26 Tuesday 24 September 2024 14:51:05 -0400 (0:00:00.080) 0:00:43.555 ***** 13731 1727203865.32430: entering _queue_task() for managed-node3/package 13731 1727203865.32968: worker is 1 (out of 1 available) 13731 1727203865.32985: exiting _queue_task() for managed-node3/package 13731 1727203865.33001: done queuing things up, now waiting for results queue to drain 13731 1727203865.33003: waiting for pending results... 13731 1727203865.33284: running TaskExecutor() for managed-node3/TASK: Install pgrep, sysctl 13731 1727203865.33397: in run() - task 028d2410-947f-82dc-c122-000000000976 13731 1727203865.33410: variable 'ansible_search_path' from source: unknown 13731 1727203865.33414: variable 'ansible_search_path' from source: unknown 13731 1727203865.33457: calling self._execute() 13731 1727203865.33552: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203865.33556: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203865.33569: variable 'omit' from source: magic vars 13731 1727203865.33928: variable 'ansible_distribution_major_version' from source: facts 13731 1727203865.33940: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203865.34058: variable 'ansible_os_family' from source: facts 13731 1727203865.34070: Evaluated conditional (ansible_os_family == 'RedHat'): True 13731 1727203865.34246: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203865.34553: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203865.34614: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203865.34643: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203865.34680: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203865.34980: variable 'ansible_distribution_major_version' from source: facts 13731 1727203865.34983: Evaluated conditional (ansible_distribution_major_version is version('7', '>=')): True 13731 1727203865.34986: variable 'omit' from source: magic vars 13731 1727203865.34988: variable 'omit' from source: magic vars 13731 1727203865.35073: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203865.37937: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203865.38024: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203865.38066: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203865.38107: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203865.38139: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203865.38243: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203865.38282: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203865.38319: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203865.38367: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203865.38389: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203865.38492: variable '__network_is_ostree' from source: set_fact 13731 1727203865.38503: variable 'omit' from source: magic vars 13731 1727203865.38540: variable 'omit' from source: magic vars 13731 1727203865.38571: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203865.38606: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203865.38635: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203865.38659: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203865.38675: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203865.38712: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203865.38721: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203865.38734: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203865.38841: Set connection var ansible_pipelining to False 13731 1727203865.38853: Set connection var ansible_shell_type to sh 13731 1727203865.38864: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203865.38873: Set connection var ansible_connection to ssh 13731 1727203865.38887: Set connection var ansible_shell_executable to /bin/sh 13731 1727203865.38897: Set connection var ansible_timeout to 10 13731 1727203865.38924: variable 'ansible_shell_executable' from source: unknown 13731 1727203865.38932: variable 'ansible_connection' from source: unknown 13731 1727203865.38946: variable 'ansible_module_compression' from source: unknown 13731 1727203865.39056: variable 'ansible_shell_type' from source: unknown 13731 1727203865.39059: variable 'ansible_shell_executable' from source: unknown 13731 1727203865.39062: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203865.39064: variable 'ansible_pipelining' from source: unknown 13731 1727203865.39066: variable 'ansible_timeout' from source: unknown 13731 1727203865.39068: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203865.39080: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203865.39095: variable 'omit' from source: magic vars 13731 1727203865.39104: starting attempt loop 13731 1727203865.39111: running the handler 13731 1727203865.39123: variable 'ansible_facts' from source: unknown 13731 1727203865.39130: variable 'ansible_facts' from source: unknown 13731 1727203865.39173: _low_level_execute_command(): starting 13731 1727203865.39187: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203865.39837: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203865.39855: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203865.39871: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203865.39985: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203865.40006: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203865.40022: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203865.40097: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203865.41965: stdout chunk (state=3): >>>/root <<< 13731 1727203865.42000: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203865.42004: stdout chunk (state=3): >>><<< 13731 1727203865.42007: stderr chunk (state=3): >>><<< 13731 1727203865.42027: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203865.42046: _low_level_execute_command(): starting 13731 1727203865.42123: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945 `" && echo ansible-tmp-1727203865.4203482-17136-279223840713945="` echo /root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945 `" ) && sleep 0' 13731 1727203865.42720: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203865.42734: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203865.42747: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203865.42762: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203865.42782: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203865.42794: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203865.42808: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203865.42901: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203865.42932: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203865.42980: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203865.43009: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203865.44868: stdout chunk (state=3): >>>ansible-tmp-1727203865.4203482-17136-279223840713945=/root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945 <<< 13731 1727203865.45014: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203865.45018: stdout chunk (state=3): >>><<< 13731 1727203865.45024: stderr chunk (state=3): >>><<< 13731 1727203865.45045: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203865.4203482-17136-279223840713945=/root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203865.45090: variable 'ansible_module_compression' from source: unknown 13731 1727203865.45154: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.dnf-ZIP_DEFLATED 13731 1727203865.45251: variable 'ansible_facts' from source: unknown 13731 1727203865.45403: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/AnsiballZ_dnf.py 13731 1727203865.45554: Sending initial data 13731 1727203865.45566: Sent initial data (152 bytes) 13731 1727203865.45977: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203865.45982: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203865.46006: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203865.46010: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203865.46012: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203865.46058: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203865.46069: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203865.46111: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203865.47656: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203865.47704: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203865.47761: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmplqlvw_ig /root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/AnsiballZ_dnf.py <<< 13731 1727203865.47764: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/AnsiballZ_dnf.py" <<< 13731 1727203865.47799: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmplqlvw_ig" to remote "/root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/AnsiballZ_dnf.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/AnsiballZ_dnf.py" <<< 13731 1727203865.48606: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203865.48734: stderr chunk (state=3): >>><<< 13731 1727203865.48736: stdout chunk (state=3): >>><<< 13731 1727203865.48738: done transferring module to remote 13731 1727203865.48739: _low_level_execute_command(): starting 13731 1727203865.48740: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/ /root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/AnsiballZ_dnf.py && sleep 0' 13731 1727203865.49118: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203865.49131: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203865.49141: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203865.49189: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203865.49201: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203865.49237: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203865.51072: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203865.51084: stdout chunk (state=3): >>><<< 13731 1727203865.51097: stderr chunk (state=3): >>><<< 13731 1727203865.51121: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203865.51129: _low_level_execute_command(): starting 13731 1727203865.51139: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/AnsiballZ_dnf.py && sleep 0' 13731 1727203865.51745: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203865.51761: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203865.51779: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203865.51796: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203865.51813: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203865.51833: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203865.51848: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203865.51868: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203865.51945: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203865.51975: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203865.52001: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203865.52016: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203865.52100: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203865.93824: stdout chunk (state=3): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["procps-ng"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} <<< 13731 1727203865.98112: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203865.98117: stdout chunk (state=3): >>><<< 13731 1727203865.98119: stderr chunk (state=3): >>><<< 13731 1727203865.98282: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["procps-ng"], "state": "present", "allow_downgrade": false, "allowerasing": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "use_backend": "auto", "best": null, "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "nobest": null, "releasever": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203865.98312: done with _execute_module (ansible.legacy.dnf, {'name': 'procps-ng', 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203865.98325: _low_level_execute_command(): starting 13731 1727203865.98351: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203865.4203482-17136-279223840713945/ > /dev/null 2>&1 && sleep 0' 13731 1727203865.99296: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203865.99311: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203865.99325: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203865.99355: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203865.99473: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203865.99583: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203865.99650: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203866.01505: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203866.01516: stdout chunk (state=3): >>><<< 13731 1727203866.01641: stderr chunk (state=3): >>><<< 13731 1727203866.01645: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203866.01648: handler run complete 13731 1727203866.01986: attempt loop complete, returning result 13731 1727203866.01990: _execute() done 13731 1727203866.01992: dumping result to json 13731 1727203866.01994: done dumping result, returning 13731 1727203866.01996: done running TaskExecutor() for managed-node3/TASK: Install pgrep, sysctl [028d2410-947f-82dc-c122-000000000976] 13731 1727203866.01998: sending task result for task 028d2410-947f-82dc-c122-000000000976 13731 1727203866.02082: done sending task result for task 028d2410-947f-82dc-c122-000000000976 13731 1727203866.02085: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 13731 1727203866.02171: no more pending results, returning what we have 13731 1727203866.02179: results queue empty 13731 1727203866.02180: checking for any_errors_fatal 13731 1727203866.02187: done checking for any_errors_fatal 13731 1727203866.02188: checking for max_fail_percentage 13731 1727203866.02190: done checking for max_fail_percentage 13731 1727203866.02191: checking to see if all hosts have failed and the running result is not ok 13731 1727203866.02192: done checking to see if all hosts have failed 13731 1727203866.02192: getting the remaining hosts for this loop 13731 1727203866.02199: done getting the remaining hosts for this loop 13731 1727203866.02203: getting the next task for host managed-node3 13731 1727203866.02212: done getting next task for host managed-node3 13731 1727203866.02214: ^ task is: TASK: Create test interfaces 13731 1727203866.02218: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203866.02222: getting variables 13731 1727203866.02224: in VariableManager get_vars() 13731 1727203866.02267: Calling all_inventory to load vars for managed-node3 13731 1727203866.02270: Calling groups_inventory to load vars for managed-node3 13731 1727203866.02273: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203866.02794: Calling all_plugins_play to load vars for managed-node3 13731 1727203866.02798: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203866.02802: Calling groups_plugins_play to load vars for managed-node3 13731 1727203866.05924: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203866.09353: done with get_vars() 13731 1727203866.09505: done getting variables 13731 1727203866.09571: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Create test interfaces] ************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:35 Tuesday 24 September 2024 14:51:06 -0400 (0:00:00.771) 0:00:44.327 ***** 13731 1727203866.09685: entering _queue_task() for managed-node3/shell 13731 1727203866.10364: worker is 1 (out of 1 available) 13731 1727203866.10492: exiting _queue_task() for managed-node3/shell 13731 1727203866.10505: done queuing things up, now waiting for results queue to drain 13731 1727203866.10507: waiting for pending results... 13731 1727203866.10873: running TaskExecutor() for managed-node3/TASK: Create test interfaces 13731 1727203866.11082: in run() - task 028d2410-947f-82dc-c122-000000000977 13731 1727203866.11481: variable 'ansible_search_path' from source: unknown 13731 1727203866.11485: variable 'ansible_search_path' from source: unknown 13731 1727203866.11490: calling self._execute() 13731 1727203866.11493: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203866.11496: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203866.11499: variable 'omit' from source: magic vars 13731 1727203866.12211: variable 'ansible_distribution_major_version' from source: facts 13731 1727203866.12230: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203866.12242: variable 'omit' from source: magic vars 13731 1727203866.12682: variable 'omit' from source: magic vars 13731 1727203866.13481: variable 'dhcp_interface1' from source: play vars 13731 1727203866.13485: variable 'dhcp_interface2' from source: play vars 13731 1727203866.13488: variable 'omit' from source: magic vars 13731 1727203866.13491: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203866.13493: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203866.13496: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203866.13498: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203866.13500: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203866.13502: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203866.13504: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203866.13506: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203866.13778: Set connection var ansible_pipelining to False 13731 1727203866.13791: Set connection var ansible_shell_type to sh 13731 1727203866.13802: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203866.13813: Set connection var ansible_connection to ssh 13731 1727203866.13824: Set connection var ansible_shell_executable to /bin/sh 13731 1727203866.13834: Set connection var ansible_timeout to 10 13731 1727203866.13866: variable 'ansible_shell_executable' from source: unknown 13731 1727203866.14282: variable 'ansible_connection' from source: unknown 13731 1727203866.14285: variable 'ansible_module_compression' from source: unknown 13731 1727203866.14288: variable 'ansible_shell_type' from source: unknown 13731 1727203866.14290: variable 'ansible_shell_executable' from source: unknown 13731 1727203866.14292: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203866.14294: variable 'ansible_pipelining' from source: unknown 13731 1727203866.14297: variable 'ansible_timeout' from source: unknown 13731 1727203866.14300: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203866.14303: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203866.14306: variable 'omit' from source: magic vars 13731 1727203866.14309: starting attempt loop 13731 1727203866.14311: running the handler 13731 1727203866.14314: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203866.14329: _low_level_execute_command(): starting 13731 1727203866.14393: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203866.15696: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203866.15712: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203866.15892: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203866.15904: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203866.15962: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203866.17630: stdout chunk (state=3): >>>/root <<< 13731 1727203866.17724: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203866.17770: stderr chunk (state=3): >>><<< 13731 1727203866.17783: stdout chunk (state=3): >>><<< 13731 1727203866.17813: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203866.17835: _low_level_execute_command(): starting 13731 1727203866.17846: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004 `" && echo ansible-tmp-1727203866.1782043-17255-55232065427004="` echo /root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004 `" ) && sleep 0' 13731 1727203866.19094: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203866.19116: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203866.19190: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203866.19288: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203866.19306: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203866.19339: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203866.21222: stdout chunk (state=3): >>>ansible-tmp-1727203866.1782043-17255-55232065427004=/root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004 <<< 13731 1727203866.21353: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203866.21394: stderr chunk (state=3): >>><<< 13731 1727203866.21403: stdout chunk (state=3): >>><<< 13731 1727203866.21428: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203866.1782043-17255-55232065427004=/root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203866.21467: variable 'ansible_module_compression' from source: unknown 13731 1727203866.21638: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203866.21960: variable 'ansible_facts' from source: unknown 13731 1727203866.21963: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/AnsiballZ_command.py 13731 1727203866.22097: Sending initial data 13731 1727203866.22290: Sent initial data (155 bytes) 13731 1727203866.23190: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203866.23203: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203866.23216: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203866.23457: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203866.23469: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203866.23494: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203866.23695: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203866.25105: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" <<< 13731 1727203866.25121: stderr chunk (state=3): >>>debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203866.25139: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203866.25165: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpn3ny6v38 /root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/AnsiballZ_command.py <<< 13731 1727203866.25190: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/AnsiballZ_command.py" <<< 13731 1727203866.25291: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpn3ny6v38" to remote "/root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/AnsiballZ_command.py" <<< 13731 1727203866.26408: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203866.26464: stderr chunk (state=3): >>><<< 13731 1727203866.26473: stdout chunk (state=3): >>><<< 13731 1727203866.26500: done transferring module to remote 13731 1727203866.26516: _low_level_execute_command(): starting 13731 1727203866.26524: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/ /root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/AnsiballZ_command.py && sleep 0' 13731 1727203866.27754: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203866.27758: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203866.27770: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203866.27887: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203866.27897: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203866.27944: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203866.28095: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203866.29774: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203866.29812: stderr chunk (state=3): >>><<< 13731 1727203866.29822: stdout chunk (state=3): >>><<< 13731 1727203866.29846: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203866.29856: _low_level_execute_command(): starting 13731 1727203866.29866: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/AnsiballZ_command.py && sleep 0' 13731 1727203866.31185: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203866.31199: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203866.31295: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203866.31330: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203866.31460: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203866.31529: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203867.68616: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ ip link add test1 type veth peer name test1p\n+ ip link add test2 type veth peer name test2p\n++ pgrep NetworkManager\n+ '[' -n 704 ']'\n+ nmcli d set test1 managed true\n+ nmcli d set test2 managed true\n+ nmcli d set test1p managed false\n+ nmcli d set test2p managed false\n+ ip link set test1p up\n+ ip link set test2p up\n+ ip link add name testbr type bridge forward_delay 0\n++ pgrep NetworkManager\n+ '[' -n 704 ']'\n+ nmcli d set testbr managed false\n+ ip link set testbr up\n+ timer=0\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ let timer+=1\n+ '[' 1 -eq 30 ']'\n+ sleep 1\n+ rc=0\n+ ip addr add 192.0.2.1/24 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip -6 addr add 2001:DB8::1/32 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ grep 'release 6' /etc/redhat-release\n+ ip link set test1p master testbr\n+ ip link set test2p master testbr\n+ systemctl is-active firewalld\ninactive\n+ dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "start": "2024-09-24 14:51:06.468825", "end": "2024-09-24 14:51:07.684218", "delta": "0:00:01.215393", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203867.70108: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203867.70115: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203867.70173: stderr chunk (state=3): >>><<< 13731 1727203867.70241: stdout chunk (state=3): >>><<< 13731 1727203867.70281: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ ip link add test1 type veth peer name test1p\n+ ip link add test2 type veth peer name test2p\n++ pgrep NetworkManager\n+ '[' -n 704 ']'\n+ nmcli d set test1 managed true\n+ nmcli d set test2 managed true\n+ nmcli d set test1p managed false\n+ nmcli d set test2p managed false\n+ ip link set test1p up\n+ ip link set test2p up\n+ ip link add name testbr type bridge forward_delay 0\n++ pgrep NetworkManager\n+ '[' -n 704 ']'\n+ nmcli d set testbr managed false\n+ ip link set testbr up\n+ timer=0\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ let timer+=1\n+ '[' 1 -eq 30 ']'\n+ sleep 1\n+ rc=0\n+ ip addr add 192.0.2.1/24 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip -6 addr add 2001:DB8::1/32 dev testbr\n+ '[' 0 '!=' 0 ']'\n+ ip addr show testbr\n+ grep -q 'inet [1-9]'\n+ grep 'release 6' /etc/redhat-release\n+ ip link set test1p master testbr\n+ ip link set test2p master testbr\n+ systemctl is-active firewalld\ninactive\n+ dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "start": "2024-09-24 14:51:06.468825", "end": "2024-09-24 14:51:07.684218", "delta": "0:00:01.215393", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203867.70385: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n "$(pgrep NetworkManager)" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the \'testbr\' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n "$(pgrep NetworkManager)" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q \'inet [1-9]\'\ndo\n let "timer+=1"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc="$?"\n if [ "$rc" != 0 ]; then\n echo NOTICE - could not add testbr - error code "$rc"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc="$?"\n if [ "$rc" != 0 ]; then\n echo NOTICE - could not add testbr - error code "$rc"\n continue\n fi\ndone\n\nif grep \'release 6\' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo \'interface testbr {\' > /etc/radvd.conf\n echo \' AdvSendAdvert on;\' >> /etc/radvd.conf\n echo \' prefix 2001:DB8::/64 { \' >> /etc/radvd.conf\n echo \' AdvOnLink on; }; \' >> /etc/radvd.conf\n echo \' }; \' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service="$service"; then\n firewall-cmd --add-service "$service"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203867.70395: _low_level_execute_command(): starting 13731 1727203867.70401: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203866.1782043-17255-55232065427004/ > /dev/null 2>&1 && sleep 0' 13731 1727203867.71681: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203867.71690: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203867.71825: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203867.71887: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203867.71938: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203867.71957: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203867.72122: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203867.73961: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203867.74028: stderr chunk (state=3): >>><<< 13731 1727203867.74032: stdout chunk (state=3): >>><<< 13731 1727203867.74183: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203867.74186: handler run complete 13731 1727203867.74188: Evaluated conditional (False): False 13731 1727203867.74190: attempt loop complete, returning result 13731 1727203867.74192: _execute() done 13731 1727203867.74194: dumping result to json 13731 1727203867.74196: done dumping result, returning 13731 1727203867.74198: done running TaskExecutor() for managed-node3/TASK: Create test interfaces [028d2410-947f-82dc-c122-000000000977] 13731 1727203867.74200: sending task result for task 028d2410-947f-82dc-c122-000000000977 13731 1727203867.74273: done sending task result for task 028d2410-947f-82dc-c122-000000000977 13731 1727203867.74279: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "set -euxo pipefail\nexec 1>&2\nip link add test1 type veth peer name test1p\nip link add test2 type veth peer name test2p\nif [ -n \"$(pgrep NetworkManager)\" ];then\n nmcli d set test1 managed true\n nmcli d set test2 managed true\n # NetworkManager should not manage DHCP server ports\n nmcli d set test1p managed false\n nmcli d set test2p managed false\nfi\nip link set test1p up\nip link set test2p up\n\n# Create the 'testbr' - providing both 10.x ipv4 and 2620:52:0 ipv6 dhcp\nip link add name testbr type bridge forward_delay 0\nif [ -n \"$(pgrep NetworkManager)\" ];then\n # NetworkManager should not manage DHCP server ports\n nmcli d set testbr managed false\nfi\nip link set testbr up\ntimer=0\n# The while loop following is a workaround for the NM bug, which can be\n# tracked in https://bugzilla.redhat.com/show_bug.cgi?id=2079642\nwhile ! ip addr show testbr | grep -q 'inet [1-9]'\ndo\n let \"timer+=1\"\n if [ $timer -eq 30 ]; then\n echo ERROR - could not add testbr\n ip addr\n exit 1\n fi\n sleep 1\n rc=0\n ip addr add 192.0.2.1/24 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\n ip -6 addr add 2001:DB8::1/32 dev testbr || rc=\"$?\"\n if [ \"$rc\" != 0 ]; then\n echo NOTICE - could not add testbr - error code \"$rc\"\n continue\n fi\ndone\n\nif grep 'release 6' /etc/redhat-release; then\n # We need bridge-utils and radvd only in rhel6\n if ! rpm -q --quiet radvd; then yum -y install radvd; fi\n if ! rpm -q --quiet bridge-utils; then yum -y install bridge-utils; fi\n\n # We need to add iptables rule to allow dhcp request\n iptables -I INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\n\n # Add test1, test2 peers into the testbr\n brctl addif testbr test1p\n brctl addif testbr test2p\n\n # in RHEL6 /run is not present\n mkdir -p /run\n\n # and dnsmasq does not support ipv6\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --interface=testbr --bind-interfaces\n\n # start radvd for ipv6\n echo 'interface testbr {' > /etc/radvd.conf\n echo ' AdvSendAdvert on;' >> /etc/radvd.conf\n echo ' prefix 2001:DB8::/64 { ' >> /etc/radvd.conf\n echo ' AdvOnLink on; }; ' >> /etc/radvd.conf\n echo ' }; ' >> /etc/radvd.conf\n\n # enable ipv6 forwarding\n sysctl -w net.ipv6.conf.all.forwarding=1\n service radvd restart\n\nelse\n ip link set test1p master testbr\n ip link set test2p master testbr\n # Run joint DHCP4/DHCP6 server with RA enabled in veth namespace\n if systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if ! firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --add-service \"$service\"\n fi\n done\n fi\n dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces\nfi\n", "delta": "0:00:01.215393", "end": "2024-09-24 14:51:07.684218", "rc": 0, "start": "2024-09-24 14:51:06.468825" } STDERR: + exec + ip link add test1 type veth peer name test1p + ip link add test2 type veth peer name test2p ++ pgrep NetworkManager + '[' -n 704 ']' + nmcli d set test1 managed true + nmcli d set test2 managed true + nmcli d set test1p managed false + nmcli d set test2p managed false + ip link set test1p up + ip link set test2p up + ip link add name testbr type bridge forward_delay 0 ++ pgrep NetworkManager + '[' -n 704 ']' + nmcli d set testbr managed false + ip link set testbr up + timer=0 + ip addr show testbr + grep -q 'inet [1-9]' + let timer+=1 + '[' 1 -eq 30 ']' + sleep 1 + rc=0 + ip addr add 192.0.2.1/24 dev testbr + '[' 0 '!=' 0 ']' + ip -6 addr add 2001:DB8::1/32 dev testbr + '[' 0 '!=' 0 ']' + ip addr show testbr + grep -q 'inet [1-9]' + grep 'release 6' /etc/redhat-release + ip link set test1p master testbr + ip link set test2p master testbr + systemctl is-active firewalld inactive + dnsmasq --pid-file=/run/dhcp_testbr.pid --dhcp-leasefile=/run/dhcp_testbr.lease --dhcp-range=192.0.2.1,192.0.2.254,240 --dhcp-range=2001:DB8::10,2001:DB8::1FF,slaac,64,240 --enable-ra --interface=testbr --bind-interfaces 13731 1727203867.74373: no more pending results, returning what we have 13731 1727203867.74380: results queue empty 13731 1727203867.74381: checking for any_errors_fatal 13731 1727203867.74392: done checking for any_errors_fatal 13731 1727203867.74393: checking for max_fail_percentage 13731 1727203867.74395: done checking for max_fail_percentage 13731 1727203867.74396: checking to see if all hosts have failed and the running result is not ok 13731 1727203867.74397: done checking to see if all hosts have failed 13731 1727203867.74397: getting the remaining hosts for this loop 13731 1727203867.74399: done getting the remaining hosts for this loop 13731 1727203867.74403: getting the next task for host managed-node3 13731 1727203867.74416: done getting next task for host managed-node3 13731 1727203867.74419: ^ task is: TASK: Include the task 'get_interface_stat.yml' 13731 1727203867.74424: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203867.74429: getting variables 13731 1727203867.74430: in VariableManager get_vars() 13731 1727203867.74472: Calling all_inventory to load vars for managed-node3 13731 1727203867.74580: Calling groups_inventory to load vars for managed-node3 13731 1727203867.74593: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203867.74605: Calling all_plugins_play to load vars for managed-node3 13731 1727203867.74608: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203867.74611: Calling groups_plugins_play to load vars for managed-node3 13731 1727203867.76453: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203867.85324: done with get_vars() 13731 1727203867.85352: done getting variables TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:3 Tuesday 24 September 2024 14:51:07 -0400 (0:00:01.757) 0:00:46.085 ***** 13731 1727203867.85448: entering _queue_task() for managed-node3/include_tasks 13731 1727203867.85949: worker is 1 (out of 1 available) 13731 1727203867.85963: exiting _queue_task() for managed-node3/include_tasks 13731 1727203867.85974: done queuing things up, now waiting for results queue to drain 13731 1727203867.85982: waiting for pending results... 13731 1727203867.86324: running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' 13731 1727203867.86344: in run() - task 028d2410-947f-82dc-c122-00000000097e 13731 1727203867.86374: variable 'ansible_search_path' from source: unknown 13731 1727203867.86420: variable 'ansible_search_path' from source: unknown 13731 1727203867.86434: calling self._execute() 13731 1727203867.86563: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203867.86586: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203867.86636: variable 'omit' from source: magic vars 13731 1727203867.87044: variable 'ansible_distribution_major_version' from source: facts 13731 1727203867.87071: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203867.87085: _execute() done 13731 1727203867.87094: dumping result to json 13731 1727203867.87119: done dumping result, returning 13731 1727203867.87124: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' [028d2410-947f-82dc-c122-00000000097e] 13731 1727203867.87132: sending task result for task 028d2410-947f-82dc-c122-00000000097e 13731 1727203867.87400: done sending task result for task 028d2410-947f-82dc-c122-00000000097e 13731 1727203867.87403: WORKER PROCESS EXITING 13731 1727203867.87433: no more pending results, returning what we have 13731 1727203867.87441: in VariableManager get_vars() 13731 1727203867.87495: Calling all_inventory to load vars for managed-node3 13731 1727203867.87498: Calling groups_inventory to load vars for managed-node3 13731 1727203867.87507: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203867.87523: Calling all_plugins_play to load vars for managed-node3 13731 1727203867.87526: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203867.87529: Calling groups_plugins_play to load vars for managed-node3 13731 1727203867.89257: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203867.90987: done with get_vars() 13731 1727203867.91010: variable 'ansible_search_path' from source: unknown 13731 1727203867.91012: variable 'ansible_search_path' from source: unknown 13731 1727203867.91050: we have included files to process 13731 1727203867.91052: generating all_blocks data 13731 1727203867.91055: done generating all_blocks data 13731 1727203867.91065: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203867.91066: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203867.91069: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203867.91278: done processing included file 13731 1727203867.91281: iterating over new_blocks loaded from include file 13731 1727203867.91282: in VariableManager get_vars() 13731 1727203867.91308: done with get_vars() 13731 1727203867.91310: filtering new block on tags 13731 1727203867.91341: done filtering new block on tags 13731 1727203867.91343: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node3 13731 1727203867.91349: extending task lists for all hosts with included blocks 13731 1727203867.91579: done extending task lists 13731 1727203867.91581: done processing included files 13731 1727203867.91581: results queue empty 13731 1727203867.91582: checking for any_errors_fatal 13731 1727203867.91589: done checking for any_errors_fatal 13731 1727203867.91590: checking for max_fail_percentage 13731 1727203867.91591: done checking for max_fail_percentage 13731 1727203867.91592: checking to see if all hosts have failed and the running result is not ok 13731 1727203867.91593: done checking to see if all hosts have failed 13731 1727203867.91594: getting the remaining hosts for this loop 13731 1727203867.91595: done getting the remaining hosts for this loop 13731 1727203867.91597: getting the next task for host managed-node3 13731 1727203867.91602: done getting next task for host managed-node3 13731 1727203867.91604: ^ task is: TASK: Get stat for interface {{ interface }} 13731 1727203867.91608: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203867.91611: getting variables 13731 1727203867.91612: in VariableManager get_vars() 13731 1727203867.91629: Calling all_inventory to load vars for managed-node3 13731 1727203867.91632: Calling groups_inventory to load vars for managed-node3 13731 1727203867.91634: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203867.91639: Calling all_plugins_play to load vars for managed-node3 13731 1727203867.91642: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203867.91644: Calling groups_plugins_play to load vars for managed-node3 13731 1727203867.92957: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203867.94550: done with get_vars() 13731 1727203867.94579: done getting variables 13731 1727203867.94750: variable 'interface' from source: task vars 13731 1727203867.94754: variable 'dhcp_interface1' from source: play vars 13731 1727203867.94825: variable 'dhcp_interface1' from source: play vars TASK [Get stat for interface test1] ******************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 Tuesday 24 September 2024 14:51:07 -0400 (0:00:00.094) 0:00:46.179 ***** 13731 1727203867.94861: entering _queue_task() for managed-node3/stat 13731 1727203867.95255: worker is 1 (out of 1 available) 13731 1727203867.95270: exiting _queue_task() for managed-node3/stat 13731 1727203867.95289: done queuing things up, now waiting for results queue to drain 13731 1727203867.95291: waiting for pending results... 13731 1727203867.95932: running TaskExecutor() for managed-node3/TASK: Get stat for interface test1 13731 1727203867.96245: in run() - task 028d2410-947f-82dc-c122-0000000009dd 13731 1727203867.96250: variable 'ansible_search_path' from source: unknown 13731 1727203867.96254: variable 'ansible_search_path' from source: unknown 13731 1727203867.96258: calling self._execute() 13731 1727203867.96352: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203867.96357: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203867.96364: variable 'omit' from source: magic vars 13731 1727203867.97282: variable 'ansible_distribution_major_version' from source: facts 13731 1727203867.97285: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203867.97288: variable 'omit' from source: magic vars 13731 1727203867.97431: variable 'omit' from source: magic vars 13731 1727203867.97883: variable 'interface' from source: task vars 13731 1727203867.97889: variable 'dhcp_interface1' from source: play vars 13731 1727203867.97892: variable 'dhcp_interface1' from source: play vars 13731 1727203867.97895: variable 'omit' from source: magic vars 13731 1727203867.97897: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203867.97899: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203867.97914: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203867.97932: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203867.97944: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203867.97976: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203867.97991: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203867.97997: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203867.98207: Set connection var ansible_pipelining to False 13731 1727203867.98211: Set connection var ansible_shell_type to sh 13731 1727203867.98214: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203867.98216: Set connection var ansible_connection to ssh 13731 1727203867.98218: Set connection var ansible_shell_executable to /bin/sh 13731 1727203867.98221: Set connection var ansible_timeout to 10 13731 1727203867.98223: variable 'ansible_shell_executable' from source: unknown 13731 1727203867.98228: variable 'ansible_connection' from source: unknown 13731 1727203867.98231: variable 'ansible_module_compression' from source: unknown 13731 1727203867.98233: variable 'ansible_shell_type' from source: unknown 13731 1727203867.98235: variable 'ansible_shell_executable' from source: unknown 13731 1727203867.98237: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203867.98239: variable 'ansible_pipelining' from source: unknown 13731 1727203867.98241: variable 'ansible_timeout' from source: unknown 13731 1727203867.98243: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203867.98582: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203867.98586: variable 'omit' from source: magic vars 13731 1727203867.98589: starting attempt loop 13731 1727203867.98591: running the handler 13731 1727203867.98593: _low_level_execute_command(): starting 13731 1727203867.98595: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203867.99515: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203867.99738: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203867.99993: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.00063: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.01736: stdout chunk (state=3): >>>/root <<< 13731 1727203868.02031: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.02035: stdout chunk (state=3): >>><<< 13731 1727203868.02038: stderr chunk (state=3): >>><<< 13731 1727203868.02041: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203868.02185: _low_level_execute_command(): starting 13731 1727203868.02189: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397 `" && echo ansible-tmp-1727203868.019656-17306-123343739455397="` echo /root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397 `" ) && sleep 0' 13731 1727203868.03349: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203868.03399: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203868.03419: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203868.03438: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203868.03468: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203868.03682: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.03805: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.03872: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.05737: stdout chunk (state=3): >>>ansible-tmp-1727203868.019656-17306-123343739455397=/root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397 <<< 13731 1727203868.06883: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.06887: stdout chunk (state=3): >>><<< 13731 1727203868.06889: stderr chunk (state=3): >>><<< 13731 1727203868.06892: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203868.019656-17306-123343739455397=/root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203868.06894: variable 'ansible_module_compression' from source: unknown 13731 1727203868.06896: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13731 1727203868.06898: variable 'ansible_facts' from source: unknown 13731 1727203868.07090: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/AnsiballZ_stat.py 13731 1727203868.07739: Sending initial data 13731 1727203868.07749: Sent initial data (152 bytes) 13731 1727203868.09130: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.09183: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203868.09393: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.09571: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.09864: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.11321: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" <<< 13731 1727203868.11325: stderr chunk (state=3): >>>debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203868.11384: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203868.11605: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpkp10ifd7 /root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/AnsiballZ_stat.py <<< 13731 1727203868.11665: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/AnsiballZ_stat.py" debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpkp10ifd7" to remote "/root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/AnsiballZ_stat.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/AnsiballZ_stat.py" <<< 13731 1727203868.12818: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.12845: stderr chunk (state=3): >>><<< 13731 1727203868.12850: stdout chunk (state=3): >>><<< 13731 1727203868.12870: done transferring module to remote 13731 1727203868.12881: _low_level_execute_command(): starting 13731 1727203868.12886: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/ /root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/AnsiballZ_stat.py && sleep 0' 13731 1727203868.13465: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203868.13469: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203868.13509: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203868.13516: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203868.13533: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.13611: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203868.13628: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.13637: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.13700: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.15546: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.15597: stderr chunk (state=3): >>><<< 13731 1727203868.15600: stdout chunk (state=3): >>><<< 13731 1727203868.15685: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203868.15689: _low_level_execute_command(): starting 13731 1727203868.15691: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/AnsiballZ_stat.py && sleep 0' 13731 1727203868.16383: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.16392: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.16456: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.31684: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test1", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 28791, "dev": 23, "nlink": 1, "atime": 1727203866.4754212, "mtime": 1727203866.4754212, "ctime": 1727203866.4754212, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test1", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13731 1727203868.33082: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203868.33094: stdout chunk (state=3): >>><<< 13731 1727203868.33103: stderr chunk (state=3): >>><<< 13731 1727203868.33124: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test1", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 28791, "dev": 23, "nlink": 1, "atime": 1727203866.4754212, "mtime": 1727203866.4754212, "ctime": 1727203866.4754212, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test1", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203868.33167: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/sys/class/net/test1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203868.33184: _low_level_execute_command(): starting 13731 1727203868.33191: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203868.019656-17306-123343739455397/ > /dev/null 2>&1 && sleep 0' 13731 1727203868.33722: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203868.33732: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203868.33744: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.33805: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203868.33812: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.33846: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.33886: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.35715: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.35742: stderr chunk (state=3): >>><<< 13731 1727203868.35744: stdout chunk (state=3): >>><<< 13731 1727203868.35769: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203868.35773: handler run complete 13731 1727203868.35808: attempt loop complete, returning result 13731 1727203868.35812: _execute() done 13731 1727203868.35814: dumping result to json 13731 1727203868.35843: done dumping result, returning 13731 1727203868.35848: done running TaskExecutor() for managed-node3/TASK: Get stat for interface test1 [028d2410-947f-82dc-c122-0000000009dd] 13731 1727203868.35867: sending task result for task 028d2410-947f-82dc-c122-0000000009dd 13731 1727203868.35993: done sending task result for task 028d2410-947f-82dc-c122-0000000009dd 13731 1727203868.35997: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "stat": { "atime": 1727203866.4754212, "block_size": 4096, "blocks": 0, "ctime": 1727203866.4754212, "dev": 23, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 28791, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": true, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/sys/devices/virtual/net/test1", "lnk_target": "../../devices/virtual/net/test1", "mode": "0777", "mtime": 1727203866.4754212, "nlink": 1, "path": "/sys/class/net/test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "wgrp": true, "woth": true, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 13731 1727203868.36169: no more pending results, returning what we have 13731 1727203868.36173: results queue empty 13731 1727203868.36174: checking for any_errors_fatal 13731 1727203868.36178: done checking for any_errors_fatal 13731 1727203868.36178: checking for max_fail_percentage 13731 1727203868.36180: done checking for max_fail_percentage 13731 1727203868.36181: checking to see if all hosts have failed and the running result is not ok 13731 1727203868.36182: done checking to see if all hosts have failed 13731 1727203868.36182: getting the remaining hosts for this loop 13731 1727203868.36184: done getting the remaining hosts for this loop 13731 1727203868.36188: getting the next task for host managed-node3 13731 1727203868.36198: done getting next task for host managed-node3 13731 1727203868.36201: ^ task is: TASK: Assert that the interface is present - '{{ interface }}' 13731 1727203868.36207: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203868.36211: getting variables 13731 1727203868.36212: in VariableManager get_vars() 13731 1727203868.36249: Calling all_inventory to load vars for managed-node3 13731 1727203868.36253: Calling groups_inventory to load vars for managed-node3 13731 1727203868.36255: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203868.36266: Calling all_plugins_play to load vars for managed-node3 13731 1727203868.36268: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203868.36271: Calling groups_plugins_play to load vars for managed-node3 13731 1727203868.37546: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203868.38788: done with get_vars() 13731 1727203868.38811: done getting variables 13731 1727203868.38857: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203868.38953: variable 'interface' from source: task vars 13731 1727203868.38956: variable 'dhcp_interface1' from source: play vars 13731 1727203868.39006: variable 'dhcp_interface1' from source: play vars TASK [Assert that the interface is present - 'test1'] ************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:5 Tuesday 24 September 2024 14:51:08 -0400 (0:00:00.441) 0:00:46.621 ***** 13731 1727203868.39032: entering _queue_task() for managed-node3/assert 13731 1727203868.39396: worker is 1 (out of 1 available) 13731 1727203868.39415: exiting _queue_task() for managed-node3/assert 13731 1727203868.39430: done queuing things up, now waiting for results queue to drain 13731 1727203868.39432: waiting for pending results... 13731 1727203868.39737: running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'test1' 13731 1727203868.39837: in run() - task 028d2410-947f-82dc-c122-00000000097f 13731 1727203868.39849: variable 'ansible_search_path' from source: unknown 13731 1727203868.39852: variable 'ansible_search_path' from source: unknown 13731 1727203868.39889: calling self._execute() 13731 1727203868.39984: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.39988: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.39999: variable 'omit' from source: magic vars 13731 1727203868.40401: variable 'ansible_distribution_major_version' from source: facts 13731 1727203868.40410: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203868.40415: variable 'omit' from source: magic vars 13731 1727203868.40485: variable 'omit' from source: magic vars 13731 1727203868.40553: variable 'interface' from source: task vars 13731 1727203868.40556: variable 'dhcp_interface1' from source: play vars 13731 1727203868.40605: variable 'dhcp_interface1' from source: play vars 13731 1727203868.40620: variable 'omit' from source: magic vars 13731 1727203868.40680: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203868.40714: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203868.40744: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203868.40763: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203868.40773: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203868.40802: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203868.40805: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.40807: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.40902: Set connection var ansible_pipelining to False 13731 1727203868.40905: Set connection var ansible_shell_type to sh 13731 1727203868.40908: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203868.40910: Set connection var ansible_connection to ssh 13731 1727203868.40933: Set connection var ansible_shell_executable to /bin/sh 13731 1727203868.40937: Set connection var ansible_timeout to 10 13731 1727203868.40939: variable 'ansible_shell_executable' from source: unknown 13731 1727203868.40942: variable 'ansible_connection' from source: unknown 13731 1727203868.40945: variable 'ansible_module_compression' from source: unknown 13731 1727203868.40949: variable 'ansible_shell_type' from source: unknown 13731 1727203868.40951: variable 'ansible_shell_executable' from source: unknown 13731 1727203868.40954: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.40958: variable 'ansible_pipelining' from source: unknown 13731 1727203868.40963: variable 'ansible_timeout' from source: unknown 13731 1727203868.40968: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.41132: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203868.41135: variable 'omit' from source: magic vars 13731 1727203868.41137: starting attempt loop 13731 1727203868.41144: running the handler 13731 1727203868.41427: variable 'interface_stat' from source: set_fact 13731 1727203868.41431: Evaluated conditional (interface_stat.stat.exists): True 13731 1727203868.41433: handler run complete 13731 1727203868.41435: attempt loop complete, returning result 13731 1727203868.41436: _execute() done 13731 1727203868.41438: dumping result to json 13731 1727203868.41440: done dumping result, returning 13731 1727203868.41441: done running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'test1' [028d2410-947f-82dc-c122-00000000097f] 13731 1727203868.41443: sending task result for task 028d2410-947f-82dc-c122-00000000097f ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203868.41622: no more pending results, returning what we have 13731 1727203868.41627: results queue empty 13731 1727203868.41628: checking for any_errors_fatal 13731 1727203868.41640: done checking for any_errors_fatal 13731 1727203868.41643: checking for max_fail_percentage 13731 1727203868.41647: done checking for max_fail_percentage 13731 1727203868.41649: checking to see if all hosts have failed and the running result is not ok 13731 1727203868.41649: done checking to see if all hosts have failed 13731 1727203868.41650: getting the remaining hosts for this loop 13731 1727203868.41652: done getting the remaining hosts for this loop 13731 1727203868.41656: getting the next task for host managed-node3 13731 1727203868.41667: done getting next task for host managed-node3 13731 1727203868.41673: ^ task is: TASK: Include the task 'get_interface_stat.yml' 13731 1727203868.41682: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203868.41686: getting variables 13731 1727203868.41688: in VariableManager get_vars() 13731 1727203868.41731: Calling all_inventory to load vars for managed-node3 13731 1727203868.41733: Calling groups_inventory to load vars for managed-node3 13731 1727203868.41736: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203868.41826: done sending task result for task 028d2410-947f-82dc-c122-00000000097f 13731 1727203868.41829: WORKER PROCESS EXITING 13731 1727203868.41890: Calling all_plugins_play to load vars for managed-node3 13731 1727203868.41894: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203868.41897: Calling groups_plugins_play to load vars for managed-node3 13731 1727203868.43303: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203868.44200: done with get_vars() 13731 1727203868.44217: done getting variables TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:3 Tuesday 24 September 2024 14:51:08 -0400 (0:00:00.053) 0:00:46.674 ***** 13731 1727203868.44346: entering _queue_task() for managed-node3/include_tasks 13731 1727203868.44657: worker is 1 (out of 1 available) 13731 1727203868.44672: exiting _queue_task() for managed-node3/include_tasks 13731 1727203868.44687: done queuing things up, now waiting for results queue to drain 13731 1727203868.44689: waiting for pending results... 13731 1727203868.44889: running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' 13731 1727203868.44980: in run() - task 028d2410-947f-82dc-c122-000000000983 13731 1727203868.44991: variable 'ansible_search_path' from source: unknown 13731 1727203868.44995: variable 'ansible_search_path' from source: unknown 13731 1727203868.45024: calling self._execute() 13731 1727203868.45111: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.45115: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.45126: variable 'omit' from source: magic vars 13731 1727203868.45416: variable 'ansible_distribution_major_version' from source: facts 13731 1727203868.45425: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203868.45431: _execute() done 13731 1727203868.45434: dumping result to json 13731 1727203868.45436: done dumping result, returning 13731 1727203868.45441: done running TaskExecutor() for managed-node3/TASK: Include the task 'get_interface_stat.yml' [028d2410-947f-82dc-c122-000000000983] 13731 1727203868.45447: sending task result for task 028d2410-947f-82dc-c122-000000000983 13731 1727203868.45532: done sending task result for task 028d2410-947f-82dc-c122-000000000983 13731 1727203868.45535: WORKER PROCESS EXITING 13731 1727203868.45589: no more pending results, returning what we have 13731 1727203868.45594: in VariableManager get_vars() 13731 1727203868.45636: Calling all_inventory to load vars for managed-node3 13731 1727203868.45639: Calling groups_inventory to load vars for managed-node3 13731 1727203868.45641: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203868.45655: Calling all_plugins_play to load vars for managed-node3 13731 1727203868.45658: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203868.45662: Calling groups_plugins_play to load vars for managed-node3 13731 1727203868.46467: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203868.47509: done with get_vars() 13731 1727203868.47525: variable 'ansible_search_path' from source: unknown 13731 1727203868.47526: variable 'ansible_search_path' from source: unknown 13731 1727203868.47550: we have included files to process 13731 1727203868.47551: generating all_blocks data 13731 1727203868.47552: done generating all_blocks data 13731 1727203868.47555: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203868.47556: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203868.47557: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml 13731 1727203868.47702: done processing included file 13731 1727203868.47703: iterating over new_blocks loaded from include file 13731 1727203868.47705: in VariableManager get_vars() 13731 1727203868.47717: done with get_vars() 13731 1727203868.47718: filtering new block on tags 13731 1727203868.47740: done filtering new block on tags 13731 1727203868.47742: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node3 13731 1727203868.47745: extending task lists for all hosts with included blocks 13731 1727203868.47870: done extending task lists 13731 1727203868.47871: done processing included files 13731 1727203868.47871: results queue empty 13731 1727203868.47872: checking for any_errors_fatal 13731 1727203868.47875: done checking for any_errors_fatal 13731 1727203868.47877: checking for max_fail_percentage 13731 1727203868.47877: done checking for max_fail_percentage 13731 1727203868.47878: checking to see if all hosts have failed and the running result is not ok 13731 1727203868.47878: done checking to see if all hosts have failed 13731 1727203868.47879: getting the remaining hosts for this loop 13731 1727203868.47880: done getting the remaining hosts for this loop 13731 1727203868.47881: getting the next task for host managed-node3 13731 1727203868.47884: done getting next task for host managed-node3 13731 1727203868.47886: ^ task is: TASK: Get stat for interface {{ interface }} 13731 1727203868.47889: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203868.47890: getting variables 13731 1727203868.47891: in VariableManager get_vars() 13731 1727203868.47899: Calling all_inventory to load vars for managed-node3 13731 1727203868.47901: Calling groups_inventory to load vars for managed-node3 13731 1727203868.47902: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203868.47906: Calling all_plugins_play to load vars for managed-node3 13731 1727203868.47907: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203868.47909: Calling groups_plugins_play to load vars for managed-node3 13731 1727203868.48815: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203868.49889: done with get_vars() 13731 1727203868.49906: done getting variables 13731 1727203868.50047: variable 'interface' from source: task vars 13731 1727203868.50050: variable 'dhcp_interface2' from source: play vars 13731 1727203868.50102: variable 'dhcp_interface2' from source: play vars TASK [Get stat for interface test2] ******************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 Tuesday 24 September 2024 14:51:08 -0400 (0:00:00.057) 0:00:46.732 ***** 13731 1727203868.50127: entering _queue_task() for managed-node3/stat 13731 1727203868.50476: worker is 1 (out of 1 available) 13731 1727203868.50491: exiting _queue_task() for managed-node3/stat 13731 1727203868.50503: done queuing things up, now waiting for results queue to drain 13731 1727203868.50505: waiting for pending results... 13731 1727203868.50739: running TaskExecutor() for managed-node3/TASK: Get stat for interface test2 13731 1727203868.50834: in run() - task 028d2410-947f-82dc-c122-000000000a01 13731 1727203868.50847: variable 'ansible_search_path' from source: unknown 13731 1727203868.50857: variable 'ansible_search_path' from source: unknown 13731 1727203868.50885: calling self._execute() 13731 1727203868.50957: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.50969: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.50983: variable 'omit' from source: magic vars 13731 1727203868.51268: variable 'ansible_distribution_major_version' from source: facts 13731 1727203868.51279: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203868.51288: variable 'omit' from source: magic vars 13731 1727203868.51350: variable 'omit' from source: magic vars 13731 1727203868.51462: variable 'interface' from source: task vars 13731 1727203868.51466: variable 'dhcp_interface2' from source: play vars 13731 1727203868.51539: variable 'dhcp_interface2' from source: play vars 13731 1727203868.51573: variable 'omit' from source: magic vars 13731 1727203868.51610: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203868.51656: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203868.51663: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203868.51687: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203868.51696: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203868.51726: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203868.51729: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.51732: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.51836: Set connection var ansible_pipelining to False 13731 1727203868.51839: Set connection var ansible_shell_type to sh 13731 1727203868.51843: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203868.51864: Set connection var ansible_connection to ssh 13731 1727203868.51867: Set connection var ansible_shell_executable to /bin/sh 13731 1727203868.51870: Set connection var ansible_timeout to 10 13731 1727203868.51879: variable 'ansible_shell_executable' from source: unknown 13731 1727203868.51882: variable 'ansible_connection' from source: unknown 13731 1727203868.51884: variable 'ansible_module_compression' from source: unknown 13731 1727203868.51887: variable 'ansible_shell_type' from source: unknown 13731 1727203868.51894: variable 'ansible_shell_executable' from source: unknown 13731 1727203868.51901: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.51907: variable 'ansible_pipelining' from source: unknown 13731 1727203868.51911: variable 'ansible_timeout' from source: unknown 13731 1727203868.51913: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.52141: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203868.52150: variable 'omit' from source: magic vars 13731 1727203868.52153: starting attempt loop 13731 1727203868.52156: running the handler 13731 1727203868.52178: _low_level_execute_command(): starting 13731 1727203868.52182: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203868.52876: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.52933: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.52938: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.52988: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.54613: stdout chunk (state=3): >>>/root <<< 13731 1727203868.54713: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.54745: stderr chunk (state=3): >>><<< 13731 1727203868.54748: stdout chunk (state=3): >>><<< 13731 1727203868.54769: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203868.54784: _low_level_execute_command(): starting 13731 1727203868.54790: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399 `" && echo ansible-tmp-1727203868.5476863-17336-58542580803399="` echo /root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399 `" ) && sleep 0' 13731 1727203868.55258: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203868.55262: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203868.55361: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.55373: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203868.55383: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.55386: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.55412: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.55467: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.57350: stdout chunk (state=3): >>>ansible-tmp-1727203868.5476863-17336-58542580803399=/root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399 <<< 13731 1727203868.57457: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.57493: stderr chunk (state=3): >>><<< 13731 1727203868.57496: stdout chunk (state=3): >>><<< 13731 1727203868.57512: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203868.5476863-17336-58542580803399=/root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203868.57554: variable 'ansible_module_compression' from source: unknown 13731 1727203868.57615: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 13731 1727203868.57652: variable 'ansible_facts' from source: unknown 13731 1727203868.57714: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/AnsiballZ_stat.py 13731 1727203868.57837: Sending initial data 13731 1727203868.57841: Sent initial data (152 bytes) 13731 1727203868.58335: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203868.58340: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203868.58343: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.58345: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203868.58351: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.58427: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203868.58432: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.58434: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.58458: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.60016: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203868.60044: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203868.60083: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmphgkqfn7u /root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/AnsiballZ_stat.py <<< 13731 1727203868.60095: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/AnsiballZ_stat.py" <<< 13731 1727203868.60133: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmphgkqfn7u" to remote "/root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/AnsiballZ_stat.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/AnsiballZ_stat.py" <<< 13731 1727203868.60672: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.60721: stderr chunk (state=3): >>><<< 13731 1727203868.60724: stdout chunk (state=3): >>><<< 13731 1727203868.60756: done transferring module to remote 13731 1727203868.60767: _low_level_execute_command(): starting 13731 1727203868.60770: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/ /root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/AnsiballZ_stat.py && sleep 0' 13731 1727203868.61226: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203868.61231: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203868.61235: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.61237: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203868.61244: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.61290: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203868.61295: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.61333: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.63041: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.63071: stderr chunk (state=3): >>><<< 13731 1727203868.63075: stdout chunk (state=3): >>><<< 13731 1727203868.63096: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203868.63103: _low_level_execute_command(): starting 13731 1727203868.63114: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/AnsiballZ_stat.py && sleep 0' 13731 1727203868.63634: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203868.63638: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.63640: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203868.63642: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.63701: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203868.63704: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.63761: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.78816: stdout chunk (state=3): >>> {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test2", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 29197, "dev": 23, "nlink": 1, "atime": 1727203866.4821606, "mtime": 1727203866.4821606, "ctime": 1727203866.4821606, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test2", "follow": false, "checksum_algorithm": "sha1"}}} <<< 13731 1727203868.80096: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203868.80100: stdout chunk (state=3): >>><<< 13731 1727203868.80102: stderr chunk (state=3): >>><<< 13731 1727203868.80124: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/sys/class/net/test2", "mode": "0777", "isdir": false, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": true, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 29197, "dev": 23, "nlink": 1, "atime": 1727203866.4821606, "mtime": 1727203866.4821606, "ctime": 1727203866.4821606, "wusr": true, "rusr": true, "xusr": true, "wgrp": true, "rgrp": true, "xgrp": true, "woth": true, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "pw_name": "root", "gr_name": "root"}, "invocation": {"module_args": {"get_attributes": false, "get_checksum": false, "get_mime": false, "path": "/sys/class/net/test2", "follow": false, "checksum_algorithm": "sha1"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203868.80201: done with _execute_module (stat, {'get_attributes': False, 'get_checksum': False, 'get_mime': False, 'path': '/sys/class/net/test2', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203868.80266: _low_level_execute_command(): starting 13731 1727203868.80271: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203868.5476863-17336-58542580803399/ > /dev/null 2>&1 && sleep 0' 13731 1727203868.80953: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203868.80968: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203868.80987: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203868.81066: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203868.81123: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203868.81141: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203868.81183: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203868.81389: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203868.83466: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203868.83470: stdout chunk (state=3): >>><<< 13731 1727203868.83472: stderr chunk (state=3): >>><<< 13731 1727203868.83474: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203868.83479: handler run complete 13731 1727203868.83481: attempt loop complete, returning result 13731 1727203868.83483: _execute() done 13731 1727203868.83485: dumping result to json 13731 1727203868.83487: done dumping result, returning 13731 1727203868.83489: done running TaskExecutor() for managed-node3/TASK: Get stat for interface test2 [028d2410-947f-82dc-c122-000000000a01] 13731 1727203868.83491: sending task result for task 028d2410-947f-82dc-c122-000000000a01 13731 1727203868.83571: done sending task result for task 028d2410-947f-82dc-c122-000000000a01 13731 1727203868.83576: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "stat": { "atime": 1727203866.4821606, "block_size": 4096, "blocks": 0, "ctime": 1727203866.4821606, "dev": 23, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 29197, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": true, "isreg": false, "issock": false, "isuid": false, "lnk_source": "/sys/devices/virtual/net/test2", "lnk_target": "../../devices/virtual/net/test2", "mode": "0777", "mtime": 1727203866.4821606, "nlink": 1, "path": "/sys/class/net/test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "wgrp": true, "woth": true, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 13731 1727203868.83681: no more pending results, returning what we have 13731 1727203868.83687: results queue empty 13731 1727203868.83687: checking for any_errors_fatal 13731 1727203868.83689: done checking for any_errors_fatal 13731 1727203868.83690: checking for max_fail_percentage 13731 1727203868.83691: done checking for max_fail_percentage 13731 1727203868.83692: checking to see if all hosts have failed and the running result is not ok 13731 1727203868.83693: done checking to see if all hosts have failed 13731 1727203868.83694: getting the remaining hosts for this loop 13731 1727203868.83696: done getting the remaining hosts for this loop 13731 1727203868.83700: getting the next task for host managed-node3 13731 1727203868.83710: done getting next task for host managed-node3 13731 1727203868.83713: ^ task is: TASK: Assert that the interface is present - '{{ interface }}' 13731 1727203868.83717: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203868.83723: getting variables 13731 1727203868.83724: in VariableManager get_vars() 13731 1727203868.83764: Calling all_inventory to load vars for managed-node3 13731 1727203868.83767: Calling groups_inventory to load vars for managed-node3 13731 1727203868.83770: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203868.84084: Calling all_plugins_play to load vars for managed-node3 13731 1727203868.84088: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203868.84092: Calling groups_plugins_play to load vars for managed-node3 13731 1727203868.85961: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203868.89232: done with get_vars() 13731 1727203868.89266: done getting variables 13731 1727203868.89381: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203868.89520: variable 'interface' from source: task vars 13731 1727203868.89524: variable 'dhcp_interface2' from source: play vars 13731 1727203868.89588: variable 'dhcp_interface2' from source: play vars TASK [Assert that the interface is present - 'test2'] ************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_present.yml:5 Tuesday 24 September 2024 14:51:08 -0400 (0:00:00.394) 0:00:47.127 ***** 13731 1727203868.89626: entering _queue_task() for managed-node3/assert 13731 1727203868.90054: worker is 1 (out of 1 available) 13731 1727203868.90066: exiting _queue_task() for managed-node3/assert 13731 1727203868.90106: done queuing things up, now waiting for results queue to drain 13731 1727203868.90108: waiting for pending results... 13731 1727203868.90499: running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'test2' 13731 1727203868.90504: in run() - task 028d2410-947f-82dc-c122-000000000984 13731 1727203868.90508: variable 'ansible_search_path' from source: unknown 13731 1727203868.90511: variable 'ansible_search_path' from source: unknown 13731 1727203868.90581: calling self._execute() 13731 1727203868.90703: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.90706: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.90709: variable 'omit' from source: magic vars 13731 1727203868.91074: variable 'ansible_distribution_major_version' from source: facts 13731 1727203868.91094: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203868.91106: variable 'omit' from source: magic vars 13731 1727203868.91183: variable 'omit' from source: magic vars 13731 1727203868.91284: variable 'interface' from source: task vars 13731 1727203868.91355: variable 'dhcp_interface2' from source: play vars 13731 1727203868.91358: variable 'dhcp_interface2' from source: play vars 13731 1727203868.91373: variable 'omit' from source: magic vars 13731 1727203868.91419: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203868.91470: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203868.91501: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203868.91525: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203868.91543: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203868.91589: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203868.91602: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.91610: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.91747: Set connection var ansible_pipelining to False 13731 1727203868.91781: Set connection var ansible_shell_type to sh 13731 1727203868.91785: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203868.91787: Set connection var ansible_connection to ssh 13731 1727203868.91797: Set connection var ansible_shell_executable to /bin/sh 13731 1727203868.91800: Set connection var ansible_timeout to 10 13731 1727203868.91870: variable 'ansible_shell_executable' from source: unknown 13731 1727203868.91907: variable 'ansible_connection' from source: unknown 13731 1727203868.91910: variable 'ansible_module_compression' from source: unknown 13731 1727203868.91912: variable 'ansible_shell_type' from source: unknown 13731 1727203868.91913: variable 'ansible_shell_executable' from source: unknown 13731 1727203868.91915: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.91917: variable 'ansible_pipelining' from source: unknown 13731 1727203868.91919: variable 'ansible_timeout' from source: unknown 13731 1727203868.91921: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.92128: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203868.92132: variable 'omit' from source: magic vars 13731 1727203868.92134: starting attempt loop 13731 1727203868.92136: running the handler 13731 1727203868.92309: variable 'interface_stat' from source: set_fact 13731 1727203868.92344: Evaluated conditional (interface_stat.stat.exists): True 13731 1727203868.92359: handler run complete 13731 1727203868.92385: attempt loop complete, returning result 13731 1727203868.92392: _execute() done 13731 1727203868.92398: dumping result to json 13731 1727203868.92454: done dumping result, returning 13731 1727203868.92457: done running TaskExecutor() for managed-node3/TASK: Assert that the interface is present - 'test2' [028d2410-947f-82dc-c122-000000000984] 13731 1727203868.92460: sending task result for task 028d2410-947f-82dc-c122-000000000984 13731 1727203868.92532: done sending task result for task 028d2410-947f-82dc-c122-000000000984 13731 1727203868.92535: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false } MSG: All assertions passed 13731 1727203868.92598: no more pending results, returning what we have 13731 1727203868.92602: results queue empty 13731 1727203868.92603: checking for any_errors_fatal 13731 1727203868.92616: done checking for any_errors_fatal 13731 1727203868.92617: checking for max_fail_percentage 13731 1727203868.92619: done checking for max_fail_percentage 13731 1727203868.92620: checking to see if all hosts have failed and the running result is not ok 13731 1727203868.92621: done checking to see if all hosts have failed 13731 1727203868.92622: getting the remaining hosts for this loop 13731 1727203868.92624: done getting the remaining hosts for this loop 13731 1727203868.92628: getting the next task for host managed-node3 13731 1727203868.92640: done getting next task for host managed-node3 13731 1727203868.92644: ^ task is: TASK: Test 13731 1727203868.92647: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203868.92653: getting variables 13731 1727203868.92655: in VariableManager get_vars() 13731 1727203868.92698: Calling all_inventory to load vars for managed-node3 13731 1727203868.92701: Calling groups_inventory to load vars for managed-node3 13731 1727203868.92704: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203868.92716: Calling all_plugins_play to load vars for managed-node3 13731 1727203868.92719: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203868.92722: Calling groups_plugins_play to load vars for managed-node3 13731 1727203868.94591: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203868.96271: done with get_vars() 13731 1727203868.96309: done getting variables TASK [Test] ******************************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:30 Tuesday 24 September 2024 14:51:08 -0400 (0:00:00.067) 0:00:47.195 ***** 13731 1727203868.96418: entering _queue_task() for managed-node3/include_tasks 13731 1727203868.96918: worker is 1 (out of 1 available) 13731 1727203868.96931: exiting _queue_task() for managed-node3/include_tasks 13731 1727203868.96998: done queuing things up, now waiting for results queue to drain 13731 1727203868.97000: waiting for pending results... 13731 1727203868.97328: running TaskExecutor() for managed-node3/TASK: Test 13731 1727203868.97380: in run() - task 028d2410-947f-82dc-c122-0000000008ee 13731 1727203868.97426: variable 'ansible_search_path' from source: unknown 13731 1727203868.97429: variable 'ansible_search_path' from source: unknown 13731 1727203868.97449: variable 'lsr_test' from source: include params 13731 1727203868.97671: variable 'lsr_test' from source: include params 13731 1727203868.97750: variable 'omit' from source: magic vars 13731 1727203868.97906: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203868.97969: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203868.97973: variable 'omit' from source: magic vars 13731 1727203868.98205: variable 'ansible_distribution_major_version' from source: facts 13731 1727203868.98220: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203868.98231: variable 'item' from source: unknown 13731 1727203868.98313: variable 'item' from source: unknown 13731 1727203868.98355: variable 'item' from source: unknown 13731 1727203868.98464: variable 'item' from source: unknown 13731 1727203868.98689: dumping result to json 13731 1727203868.98692: done dumping result, returning 13731 1727203868.98695: done running TaskExecutor() for managed-node3/TASK: Test [028d2410-947f-82dc-c122-0000000008ee] 13731 1727203868.98697: sending task result for task 028d2410-947f-82dc-c122-0000000008ee 13731 1727203868.98823: no more pending results, returning what we have 13731 1727203868.98831: in VariableManager get_vars() 13731 1727203868.98879: Calling all_inventory to load vars for managed-node3 13731 1727203868.98883: Calling groups_inventory to load vars for managed-node3 13731 1727203868.98886: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203868.98901: Calling all_plugins_play to load vars for managed-node3 13731 1727203868.98903: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203868.98906: Calling groups_plugins_play to load vars for managed-node3 13731 1727203868.99508: done sending task result for task 028d2410-947f-82dc-c122-0000000008ee 13731 1727203868.99512: WORKER PROCESS EXITING 13731 1727203869.00993: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203869.02729: done with get_vars() 13731 1727203869.02756: variable 'ansible_search_path' from source: unknown 13731 1727203869.02757: variable 'ansible_search_path' from source: unknown 13731 1727203869.02810: we have included files to process 13731 1727203869.02812: generating all_blocks data 13731 1727203869.02814: done generating all_blocks data 13731 1727203869.02819: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile_reconfigure.yml 13731 1727203869.02820: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile_reconfigure.yml 13731 1727203869.02823: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile_reconfigure.yml 13731 1727203869.03073: in VariableManager get_vars() 13731 1727203869.03098: done with get_vars() 13731 1727203869.03102: variable 'omit' from source: magic vars 13731 1727203869.03145: variable 'omit' from source: magic vars 13731 1727203869.03204: in VariableManager get_vars() 13731 1727203869.03219: done with get_vars() 13731 1727203869.03252: in VariableManager get_vars() 13731 1727203869.03271: done with get_vars() 13731 1727203869.03308: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/defaults/main.yml 13731 1727203869.03498: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/meta/main.yml 13731 1727203869.03583: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml 13731 1727203869.03979: in VariableManager get_vars() 13731 1727203869.04009: done with get_vars() redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13731 1727203869.06110: done processing included file 13731 1727203869.06113: iterating over new_blocks loaded from include file 13731 1727203869.06114: in VariableManager get_vars() 13731 1727203869.06147: done with get_vars() 13731 1727203869.06149: filtering new block on tags 13731 1727203869.06479: done filtering new block on tags 13731 1727203869.06484: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile_reconfigure.yml for managed-node3 => (item=tasks/create_bond_profile_reconfigure.yml) 13731 1727203869.06489: extending task lists for all hosts with included blocks 13731 1727203869.07797: done extending task lists 13731 1727203869.07800: done processing included files 13731 1727203869.07807: results queue empty 13731 1727203869.07808: checking for any_errors_fatal 13731 1727203869.07811: done checking for any_errors_fatal 13731 1727203869.07812: checking for max_fail_percentage 13731 1727203869.07813: done checking for max_fail_percentage 13731 1727203869.07814: checking to see if all hosts have failed and the running result is not ok 13731 1727203869.07815: done checking to see if all hosts have failed 13731 1727203869.07815: getting the remaining hosts for this loop 13731 1727203869.07817: done getting the remaining hosts for this loop 13731 1727203869.07819: getting the next task for host managed-node3 13731 1727203869.07824: done getting next task for host managed-node3 13731 1727203869.07827: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13731 1727203869.07831: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203869.07842: getting variables 13731 1727203869.07843: in VariableManager get_vars() 13731 1727203869.07863: Calling all_inventory to load vars for managed-node3 13731 1727203869.07865: Calling groups_inventory to load vars for managed-node3 13731 1727203869.07867: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203869.07873: Calling all_plugins_play to load vars for managed-node3 13731 1727203869.07877: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203869.07879: Calling groups_plugins_play to load vars for managed-node3 13731 1727203869.09238: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203869.11521: done with get_vars() 13731 1727203869.11553: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:4 Tuesday 24 September 2024 14:51:09 -0400 (0:00:00.153) 0:00:47.348 ***** 13731 1727203869.11767: entering _queue_task() for managed-node3/include_tasks 13731 1727203869.12570: worker is 1 (out of 1 available) 13731 1727203869.12584: exiting _queue_task() for managed-node3/include_tasks 13731 1727203869.12596: done queuing things up, now waiting for results queue to drain 13731 1727203869.12598: waiting for pending results... 13731 1727203869.13186: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13731 1727203869.13544: in run() - task 028d2410-947f-82dc-c122-000000000a2e 13731 1727203869.13559: variable 'ansible_search_path' from source: unknown 13731 1727203869.13562: variable 'ansible_search_path' from source: unknown 13731 1727203869.13601: calling self._execute() 13731 1727203869.13899: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203869.13903: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203869.13913: variable 'omit' from source: magic vars 13731 1727203869.14694: variable 'ansible_distribution_major_version' from source: facts 13731 1727203869.14704: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203869.14716: _execute() done 13731 1727203869.14720: dumping result to json 13731 1727203869.14723: done dumping result, returning 13731 1727203869.14726: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role [028d2410-947f-82dc-c122-000000000a2e] 13731 1727203869.14728: sending task result for task 028d2410-947f-82dc-c122-000000000a2e 13731 1727203869.14940: no more pending results, returning what we have 13731 1727203869.14947: in VariableManager get_vars() 13731 1727203869.15154: Calling all_inventory to load vars for managed-node3 13731 1727203869.15157: Calling groups_inventory to load vars for managed-node3 13731 1727203869.15159: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203869.15172: Calling all_plugins_play to load vars for managed-node3 13731 1727203869.15177: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203869.15179: Calling groups_plugins_play to load vars for managed-node3 13731 1727203869.15794: done sending task result for task 028d2410-947f-82dc-c122-000000000a2e 13731 1727203869.15798: WORKER PROCESS EXITING 13731 1727203869.17952: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203869.20995: done with get_vars() 13731 1727203869.21028: variable 'ansible_search_path' from source: unknown 13731 1727203869.21029: variable 'ansible_search_path' from source: unknown 13731 1727203869.21072: we have included files to process 13731 1727203869.21195: generating all_blocks data 13731 1727203869.21198: done generating all_blocks data 13731 1727203869.21199: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203869.21201: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203869.21204: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203869.22288: done processing included file 13731 1727203869.22290: iterating over new_blocks loaded from include file 13731 1727203869.22292: in VariableManager get_vars() 13731 1727203869.22323: done with get_vars() 13731 1727203869.22326: filtering new block on tags 13731 1727203869.22366: done filtering new block on tags 13731 1727203869.22370: in VariableManager get_vars() 13731 1727203869.22420: done with get_vars() 13731 1727203869.22422: filtering new block on tags 13731 1727203869.22480: done filtering new block on tags 13731 1727203869.22483: in VariableManager get_vars() 13731 1727203869.22510: done with get_vars() 13731 1727203869.22512: filtering new block on tags 13731 1727203869.22564: done filtering new block on tags 13731 1727203869.22567: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml for managed-node3 13731 1727203869.22572: extending task lists for all hosts with included blocks 13731 1727203869.24485: done extending task lists 13731 1727203869.24487: done processing included files 13731 1727203869.24488: results queue empty 13731 1727203869.24488: checking for any_errors_fatal 13731 1727203869.24501: done checking for any_errors_fatal 13731 1727203869.24529: checking for max_fail_percentage 13731 1727203869.24530: done checking for max_fail_percentage 13731 1727203869.24532: checking to see if all hosts have failed and the running result is not ok 13731 1727203869.24532: done checking to see if all hosts have failed 13731 1727203869.24533: getting the remaining hosts for this loop 13731 1727203869.24534: done getting the remaining hosts for this loop 13731 1727203869.24537: getting the next task for host managed-node3 13731 1727203869.24542: done getting next task for host managed-node3 13731 1727203869.24545: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13731 1727203869.24549: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203869.24559: getting variables 13731 1727203869.24561: in VariableManager get_vars() 13731 1727203869.24606: Calling all_inventory to load vars for managed-node3 13731 1727203869.24609: Calling groups_inventory to load vars for managed-node3 13731 1727203869.24611: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203869.24617: Calling all_plugins_play to load vars for managed-node3 13731 1727203869.24643: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203869.24657: Calling groups_plugins_play to load vars for managed-node3 13731 1727203869.25940: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203869.27707: done with get_vars() 13731 1727203869.27733: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role are present] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:3 Tuesday 24 September 2024 14:51:09 -0400 (0:00:00.160) 0:00:47.509 ***** 13731 1727203869.27824: entering _queue_task() for managed-node3/setup 13731 1727203869.28197: worker is 1 (out of 1 available) 13731 1727203869.28386: exiting _queue_task() for managed-node3/setup 13731 1727203869.28399: done queuing things up, now waiting for results queue to drain 13731 1727203869.28419: waiting for pending results... 13731 1727203869.28553: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13731 1727203869.28728: in run() - task 028d2410-947f-82dc-c122-000000000b10 13731 1727203869.28752: variable 'ansible_search_path' from source: unknown 13731 1727203869.28762: variable 'ansible_search_path' from source: unknown 13731 1727203869.28816: calling self._execute() 13731 1727203869.28924: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203869.28935: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203869.28948: variable 'omit' from source: magic vars 13731 1727203869.29356: variable 'ansible_distribution_major_version' from source: facts 13731 1727203869.29373: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203869.29628: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203869.32772: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203869.32823: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203869.32870: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203869.32917: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203869.32953: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203869.33083: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203869.33097: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203869.33124: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203869.33170: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203869.33258: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203869.33265: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203869.33295: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203869.33322: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203869.33359: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203869.33384: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203869.33555: variable '__network_required_facts' from source: role '' defaults 13731 1727203869.33570: variable 'ansible_facts' from source: unknown 13731 1727203869.34370: Evaluated conditional (__network_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 13731 1727203869.34386: when evaluation is False, skipping this task 13731 1727203869.34394: _execute() done 13731 1727203869.34453: dumping result to json 13731 1727203869.34456: done dumping result, returning 13731 1727203869.34459: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present [028d2410-947f-82dc-c122-000000000b10] 13731 1727203869.34462: sending task result for task 028d2410-947f-82dc-c122-000000000b10 skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203869.34608: no more pending results, returning what we have 13731 1727203869.34613: results queue empty 13731 1727203869.34614: checking for any_errors_fatal 13731 1727203869.34616: done checking for any_errors_fatal 13731 1727203869.34616: checking for max_fail_percentage 13731 1727203869.34618: done checking for max_fail_percentage 13731 1727203869.34619: checking to see if all hosts have failed and the running result is not ok 13731 1727203869.34620: done checking to see if all hosts have failed 13731 1727203869.34621: getting the remaining hosts for this loop 13731 1727203869.34623: done getting the remaining hosts for this loop 13731 1727203869.34627: getting the next task for host managed-node3 13731 1727203869.34640: done getting next task for host managed-node3 13731 1727203869.34644: ^ task is: TASK: fedora.linux_system_roles.network : Check if system is ostree 13731 1727203869.34650: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203869.34682: getting variables 13731 1727203869.34685: in VariableManager get_vars() 13731 1727203869.34731: Calling all_inventory to load vars for managed-node3 13731 1727203869.34734: Calling groups_inventory to load vars for managed-node3 13731 1727203869.34737: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203869.34749: Calling all_plugins_play to load vars for managed-node3 13731 1727203869.34753: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203869.34756: Calling groups_plugins_play to load vars for managed-node3 13731 1727203869.35391: done sending task result for task 028d2410-947f-82dc-c122-000000000b10 13731 1727203869.35401: WORKER PROCESS EXITING 13731 1727203869.36864: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203869.40155: done with get_vars() 13731 1727203869.40220: done getting variables TASK [fedora.linux_system_roles.network : Check if system is ostree] *********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:12 Tuesday 24 September 2024 14:51:09 -0400 (0:00:00.125) 0:00:47.634 ***** 13731 1727203869.40356: entering _queue_task() for managed-node3/stat 13731 1727203869.40973: worker is 1 (out of 1 available) 13731 1727203869.41020: exiting _queue_task() for managed-node3/stat 13731 1727203869.41031: done queuing things up, now waiting for results queue to drain 13731 1727203869.41033: waiting for pending results... 13731 1727203869.41287: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if system is ostree 13731 1727203869.41746: in run() - task 028d2410-947f-82dc-c122-000000000b12 13731 1727203869.41878: variable 'ansible_search_path' from source: unknown 13731 1727203869.41882: variable 'ansible_search_path' from source: unknown 13731 1727203869.41886: calling self._execute() 13731 1727203869.42087: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203869.42092: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203869.42098: variable 'omit' from source: magic vars 13731 1727203869.42781: variable 'ansible_distribution_major_version' from source: facts 13731 1727203869.42785: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203869.43209: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203869.43610: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203869.43890: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203869.43894: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203869.43944: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203869.44051: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203869.44092: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203869.44129: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203869.44164: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203869.44299: variable '__network_is_ostree' from source: set_fact 13731 1727203869.44303: Evaluated conditional (not __network_is_ostree is defined): False 13731 1727203869.44305: when evaluation is False, skipping this task 13731 1727203869.44308: _execute() done 13731 1727203869.44310: dumping result to json 13731 1727203869.44317: done dumping result, returning 13731 1727203869.44322: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if system is ostree [028d2410-947f-82dc-c122-000000000b12] 13731 1727203869.44330: sending task result for task 028d2410-947f-82dc-c122-000000000b12 13731 1727203869.44487: done sending task result for task 028d2410-947f-82dc-c122-000000000b12 13731 1727203869.44491: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13731 1727203869.44706: no more pending results, returning what we have 13731 1727203869.44711: results queue empty 13731 1727203869.44712: checking for any_errors_fatal 13731 1727203869.44719: done checking for any_errors_fatal 13731 1727203869.44720: checking for max_fail_percentage 13731 1727203869.44721: done checking for max_fail_percentage 13731 1727203869.44722: checking to see if all hosts have failed and the running result is not ok 13731 1727203869.44723: done checking to see if all hosts have failed 13731 1727203869.44724: getting the remaining hosts for this loop 13731 1727203869.44726: done getting the remaining hosts for this loop 13731 1727203869.44729: getting the next task for host managed-node3 13731 1727203869.44737: done getting next task for host managed-node3 13731 1727203869.44741: ^ task is: TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13731 1727203869.44747: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203869.44771: getting variables 13731 1727203869.44773: in VariableManager get_vars() 13731 1727203869.44821: Calling all_inventory to load vars for managed-node3 13731 1727203869.44824: Calling groups_inventory to load vars for managed-node3 13731 1727203869.44827: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203869.44836: Calling all_plugins_play to load vars for managed-node3 13731 1727203869.44839: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203869.44842: Calling groups_plugins_play to load vars for managed-node3 13731 1727203869.46408: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203869.48540: done with get_vars() 13731 1727203869.48603: done getting variables 13731 1727203869.48667: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Set flag to indicate system is ostree] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:17 Tuesday 24 September 2024 14:51:09 -0400 (0:00:00.083) 0:00:47.718 ***** 13731 1727203869.48742: entering _queue_task() for managed-node3/set_fact 13731 1727203869.49438: worker is 1 (out of 1 available) 13731 1727203869.49513: exiting _queue_task() for managed-node3/set_fact 13731 1727203869.49554: done queuing things up, now waiting for results queue to drain 13731 1727203869.49556: waiting for pending results... 13731 1727203869.50194: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13731 1727203869.50430: in run() - task 028d2410-947f-82dc-c122-000000000b13 13731 1727203869.50537: variable 'ansible_search_path' from source: unknown 13731 1727203869.50541: variable 'ansible_search_path' from source: unknown 13731 1727203869.50545: calling self._execute() 13731 1727203869.50805: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203869.50808: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203869.50811: variable 'omit' from source: magic vars 13731 1727203869.51610: variable 'ansible_distribution_major_version' from source: facts 13731 1727203869.51629: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203869.52065: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203869.52572: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203869.52630: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203869.52684: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203869.52753: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203869.52877: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203869.52918: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203869.52948: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203869.52992: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203869.53101: variable '__network_is_ostree' from source: set_fact 13731 1727203869.53132: Evaluated conditional (not __network_is_ostree is defined): False 13731 1727203869.53135: when evaluation is False, skipping this task 13731 1727203869.53138: _execute() done 13731 1727203869.53139: dumping result to json 13731 1727203869.53142: done dumping result, returning 13731 1727203869.53152: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree [028d2410-947f-82dc-c122-000000000b13] 13731 1727203869.53210: sending task result for task 028d2410-947f-82dc-c122-000000000b13 13731 1727203869.53289: done sending task result for task 028d2410-947f-82dc-c122-000000000b13 13731 1727203869.53293: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13731 1727203869.53370: no more pending results, returning what we have 13731 1727203869.53374: results queue empty 13731 1727203869.53377: checking for any_errors_fatal 13731 1727203869.53384: done checking for any_errors_fatal 13731 1727203869.53384: checking for max_fail_percentage 13731 1727203869.53386: done checking for max_fail_percentage 13731 1727203869.53387: checking to see if all hosts have failed and the running result is not ok 13731 1727203869.53388: done checking to see if all hosts have failed 13731 1727203869.53389: getting the remaining hosts for this loop 13731 1727203869.53391: done getting the remaining hosts for this loop 13731 1727203869.53395: getting the next task for host managed-node3 13731 1727203869.53410: done getting next task for host managed-node3 13731 1727203869.53413: ^ task is: TASK: fedora.linux_system_roles.network : Check which services are running 13731 1727203869.53420: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203869.53443: getting variables 13731 1727203869.53445: in VariableManager get_vars() 13731 1727203869.53705: Calling all_inventory to load vars for managed-node3 13731 1727203869.53708: Calling groups_inventory to load vars for managed-node3 13731 1727203869.53710: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203869.53720: Calling all_plugins_play to load vars for managed-node3 13731 1727203869.53723: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203869.53725: Calling groups_plugins_play to load vars for managed-node3 13731 1727203869.55972: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203869.57927: done with get_vars() 13731 1727203869.58013: done getting variables TASK [fedora.linux_system_roles.network : Check which services are running] **** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 Tuesday 24 September 2024 14:51:09 -0400 (0:00:00.094) 0:00:47.812 ***** 13731 1727203869.58157: entering _queue_task() for managed-node3/service_facts 13731 1727203869.58736: worker is 1 (out of 1 available) 13731 1727203869.58748: exiting _queue_task() for managed-node3/service_facts 13731 1727203869.58763: done queuing things up, now waiting for results queue to drain 13731 1727203869.58764: waiting for pending results... 13731 1727203869.59214: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which services are running 13731 1727203869.59565: in run() - task 028d2410-947f-82dc-c122-000000000b15 13731 1727203869.59606: variable 'ansible_search_path' from source: unknown 13731 1727203869.59611: variable 'ansible_search_path' from source: unknown 13731 1727203869.59634: calling self._execute() 13731 1727203869.59826: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203869.59830: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203869.59832: variable 'omit' from source: magic vars 13731 1727203869.60463: variable 'ansible_distribution_major_version' from source: facts 13731 1727203869.60598: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203869.60626: variable 'omit' from source: magic vars 13731 1727203869.60801: variable 'omit' from source: magic vars 13731 1727203869.60805: variable 'omit' from source: magic vars 13731 1727203869.60854: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203869.60982: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203869.61011: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203869.61038: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203869.61056: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203869.61249: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203869.61257: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203869.61263: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203869.61570: Set connection var ansible_pipelining to False 13731 1727203869.61573: Set connection var ansible_shell_type to sh 13731 1727203869.61577: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203869.61580: Set connection var ansible_connection to ssh 13731 1727203869.61582: Set connection var ansible_shell_executable to /bin/sh 13731 1727203869.61584: Set connection var ansible_timeout to 10 13731 1727203869.61590: variable 'ansible_shell_executable' from source: unknown 13731 1727203869.61593: variable 'ansible_connection' from source: unknown 13731 1727203869.61595: variable 'ansible_module_compression' from source: unknown 13731 1727203869.61597: variable 'ansible_shell_type' from source: unknown 13731 1727203869.61599: variable 'ansible_shell_executable' from source: unknown 13731 1727203869.61601: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203869.61603: variable 'ansible_pipelining' from source: unknown 13731 1727203869.61605: variable 'ansible_timeout' from source: unknown 13731 1727203869.61607: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203869.61933: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203869.61940: variable 'omit' from source: magic vars 13731 1727203869.61942: starting attempt loop 13731 1727203869.61944: running the handler 13731 1727203869.62019: _low_level_execute_command(): starting 13731 1727203869.62021: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203869.62934: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203869.62952: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203869.63040: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203869.63052: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203869.63098: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203869.63227: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203869.64893: stdout chunk (state=3): >>>/root <<< 13731 1727203869.65033: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203869.65052: stderr chunk (state=3): >>><<< 13731 1727203869.65066: stdout chunk (state=3): >>><<< 13731 1727203869.65208: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203869.65211: _low_level_execute_command(): starting 13731 1727203869.65215: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406 `" && echo ansible-tmp-1727203869.6509788-17373-271087377396406="` echo /root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406 `" ) && sleep 0' 13731 1727203869.65830: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203869.65898: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203869.65917: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203869.66201: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203869.66255: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203869.68119: stdout chunk (state=3): >>>ansible-tmp-1727203869.6509788-17373-271087377396406=/root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406 <<< 13731 1727203869.68280: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203869.68293: stdout chunk (state=3): >>><<< 13731 1727203869.68309: stderr chunk (state=3): >>><<< 13731 1727203869.68340: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203869.6509788-17373-271087377396406=/root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203869.68406: variable 'ansible_module_compression' from source: unknown 13731 1727203869.68484: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.service_facts-ZIP_DEFLATED 13731 1727203869.68544: variable 'ansible_facts' from source: unknown 13731 1727203869.68648: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/AnsiballZ_service_facts.py 13731 1727203869.68878: Sending initial data 13731 1727203869.68890: Sent initial data (162 bytes) 13731 1727203869.69501: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203869.69546: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203869.69627: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203869.69630: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203869.69661: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203869.69803: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203869.69952: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203869.71554: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203869.71582: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203869.71667: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpw8rhbm_3 /root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/AnsiballZ_service_facts.py <<< 13731 1727203869.71670: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/AnsiballZ_service_facts.py" debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpw8rhbm_3" to remote "/root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/AnsiballZ_service_facts.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/AnsiballZ_service_facts.py" <<< 13731 1727203869.73849: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203869.73854: stdout chunk (state=3): >>><<< 13731 1727203869.73857: stderr chunk (state=3): >>><<< 13731 1727203869.73862: done transferring module to remote 13731 1727203869.73864: _low_level_execute_command(): starting 13731 1727203869.73867: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/ /root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/AnsiballZ_service_facts.py && sleep 0' 13731 1727203869.75714: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203869.75805: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203869.75826: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203869.75943: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203869.77743: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203869.77765: stdout chunk (state=3): >>><<< 13731 1727203869.77793: stderr chunk (state=3): >>><<< 13731 1727203869.78005: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203869.78009: _low_level_execute_command(): starting 13731 1727203869.78012: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/AnsiballZ_service_facts.py && sleep 0' 13731 1727203869.79315: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203869.79599: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203869.79893: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203869.80016: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203869.80167: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203871.29137: stdout chunk (state=3): >>> {"ansible_facts": {"services": {"audit-rules.service": {"name": "audit-rules.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dm-event.service": {"name": "dm-event.service", "state": "stopped", "status": "static", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-lvmpolld.service": {"name": "lvm2-lvmpolld.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-monitor.service": {"name": "lvm2-monitor.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@dm_mod.service": {"name": "modprobe@dm_mod.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@efi_pstore.service": {"name": "modprobe@efi_pstore.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@loop.service": {"name": "modprobe@loop.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "running", "status": "enabled", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "pcscd.service": {"name": "pcscd.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ssh-host-keys-migration.service": {"name": "ssh-host-keys-migration.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-battery-check.service": {"name": "systemd-battery-check.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-confext.service": {"name": "systemd-confext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-clear.service": {"name": "systemd-hibernate-clear.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-resume.service": {"name": "systemd-hibernate-resume.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-oomd.service": {"name": "systemd-oomd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-quotacheck-root.service": {"name": "systemd-quotacheck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-soft-reboot.service": {"name": "systemd-soft-reboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev-early.service": {"name": "systemd-tmpfiles-setup-dev-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup-early.service": {"name": "systemd-tpm2-setup-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup.service": {"name": "systemd-tpm2-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-load-credentials.service": {"name": "systemd-udev-load-credentials.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "blk-availability.service": {"name": "blk-availability.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "capsule@.service": {"name": "capsule@.service", "state": "unknown", "status": "static", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd.service": {"name": "dhcpcd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd@.service": {"name": "dhcpcd@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fips-crypto-policy-overlay.service": {"name": "fips-crypto-policy-overlay.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fsidd.service": {"name": "fsidd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "lvm-devices-import.service": {"name": "lvm-devices-import.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "polkit.service": {"name": "polkit.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon-root.service": {"name": "quotaon-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "quotaon@.service": {"name": "quotaon@.service", "state": "unknown", "status": "static", "source": "systemd"}, "rpmdb-migrate.service": {"name": "rpmdb-migrate.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-bootctl@.service": {"name": "systemd-bootctl@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-creds@.service": {"name": "systemd-creds@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald-sync@.service": {"name": "systemd-journald-sync@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrextend@.service": {"name": "systemd-pcrextend@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrlock-file-system.service": {"name": "systemd-pcrlock-file-system.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-code.service": {"name": "systemd-pcrlock-firmware-code.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-config.service": {"name": "systemd-pcrlock-firmware-config.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-machine-id.service": {"name": "systemd-pcrlock-machine-id.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-make-policy.service": {"name": "systemd-pcrlock-make-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-authority.service": {"name": "systemd-pcrlock-secureboot-authority.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-policy.service": {"name": "systemd-pcrlock-secureboot-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock@.service": {"name": "systemd-pcrlock@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-quotacheck@.service": {"name": "systemd-quotacheck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysext@.service": {"name": "systemd-sysext@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} <<< 13731 1727203871.30722: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203871.30736: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203871.31181: stderr chunk (state=3): >>><<< 13731 1727203871.31185: stdout chunk (state=3): >>><<< 13731 1727203871.31190: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"services": {"audit-rules.service": {"name": "audit-rules.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dm-event.service": {"name": "dm-event.service", "state": "stopped", "status": "static", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-lvmpolld.service": {"name": "lvm2-lvmpolld.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-monitor.service": {"name": "lvm2-monitor.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@dm_mod.service": {"name": "modprobe@dm_mod.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@efi_pstore.service": {"name": "modprobe@efi_pstore.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@loop.service": {"name": "modprobe@loop.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "running", "status": "enabled", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "pcscd.service": {"name": "pcscd.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ssh-host-keys-migration.service": {"name": "ssh-host-keys-migration.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-battery-check.service": {"name": "systemd-battery-check.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-confext.service": {"name": "systemd-confext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-clear.service": {"name": "systemd-hibernate-clear.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-resume.service": {"name": "systemd-hibernate-resume.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-oomd.service": {"name": "systemd-oomd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-quotacheck-root.service": {"name": "systemd-quotacheck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-soft-reboot.service": {"name": "systemd-soft-reboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev-early.service": {"name": "systemd-tmpfiles-setup-dev-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup-early.service": {"name": "systemd-tpm2-setup-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup.service": {"name": "systemd-tpm2-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-load-credentials.service": {"name": "systemd-udev-load-credentials.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "blk-availability.service": {"name": "blk-availability.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "capsule@.service": {"name": "capsule@.service", "state": "unknown", "status": "static", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd.service": {"name": "dhcpcd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd@.service": {"name": "dhcpcd@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fips-crypto-policy-overlay.service": {"name": "fips-crypto-policy-overlay.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fsidd.service": {"name": "fsidd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "lvm-devices-import.service": {"name": "lvm-devices-import.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "polkit.service": {"name": "polkit.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon-root.service": {"name": "quotaon-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "quotaon@.service": {"name": "quotaon@.service", "state": "unknown", "status": "static", "source": "systemd"}, "rpmdb-migrate.service": {"name": "rpmdb-migrate.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-bootctl@.service": {"name": "systemd-bootctl@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-creds@.service": {"name": "systemd-creds@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald-sync@.service": {"name": "systemd-journald-sync@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrextend@.service": {"name": "systemd-pcrextend@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrlock-file-system.service": {"name": "systemd-pcrlock-file-system.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-code.service": {"name": "systemd-pcrlock-firmware-code.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-config.service": {"name": "systemd-pcrlock-firmware-config.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-machine-id.service": {"name": "systemd-pcrlock-machine-id.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-make-policy.service": {"name": "systemd-pcrlock-make-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-authority.service": {"name": "systemd-pcrlock-secureboot-authority.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-policy.service": {"name": "systemd-pcrlock-secureboot-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock@.service": {"name": "systemd-pcrlock@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-quotacheck@.service": {"name": "systemd-quotacheck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysext@.service": {"name": "systemd-sysext@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203871.31933: done with _execute_module (service_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'service_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203871.31949: _low_level_execute_command(): starting 13731 1727203871.31959: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203869.6509788-17373-271087377396406/ > /dev/null 2>&1 && sleep 0' 13731 1727203871.32561: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203871.32583: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203871.32692: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203871.32707: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203871.32762: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203871.34752: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203871.34765: stdout chunk (state=3): >>><<< 13731 1727203871.34785: stderr chunk (state=3): >>><<< 13731 1727203871.34806: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203871.34818: handler run complete 13731 1727203871.35316: variable 'ansible_facts' from source: unknown 13731 1727203871.35562: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203871.36469: variable 'ansible_facts' from source: unknown 13731 1727203871.36714: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203871.36924: attempt loop complete, returning result 13731 1727203871.36927: _execute() done 13731 1727203871.36930: dumping result to json 13731 1727203871.37026: done dumping result, returning 13731 1727203871.37030: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which services are running [028d2410-947f-82dc-c122-000000000b15] 13731 1727203871.37032: sending task result for task 028d2410-947f-82dc-c122-000000000b15 13731 1727203871.38560: done sending task result for task 028d2410-947f-82dc-c122-000000000b15 13731 1727203871.38564: WORKER PROCESS EXITING ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203871.38666: no more pending results, returning what we have 13731 1727203871.38669: results queue empty 13731 1727203871.38670: checking for any_errors_fatal 13731 1727203871.38675: done checking for any_errors_fatal 13731 1727203871.38678: checking for max_fail_percentage 13731 1727203871.38680: done checking for max_fail_percentage 13731 1727203871.38681: checking to see if all hosts have failed and the running result is not ok 13731 1727203871.38682: done checking to see if all hosts have failed 13731 1727203871.38682: getting the remaining hosts for this loop 13731 1727203871.38684: done getting the remaining hosts for this loop 13731 1727203871.38688: getting the next task for host managed-node3 13731 1727203871.38695: done getting next task for host managed-node3 13731 1727203871.38699: ^ task is: TASK: fedora.linux_system_roles.network : Check which packages are installed 13731 1727203871.38782: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203871.38795: getting variables 13731 1727203871.38797: in VariableManager get_vars() 13731 1727203871.38844: Calling all_inventory to load vars for managed-node3 13731 1727203871.38848: Calling groups_inventory to load vars for managed-node3 13731 1727203871.38855: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203871.38867: Calling all_plugins_play to load vars for managed-node3 13731 1727203871.38870: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203871.38873: Calling groups_plugins_play to load vars for managed-node3 13731 1727203871.40433: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203871.42835: done with get_vars() 13731 1727203871.42870: done getting variables TASK [fedora.linux_system_roles.network : Check which packages are installed] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 Tuesday 24 September 2024 14:51:11 -0400 (0:00:01.848) 0:00:49.660 ***** 13731 1727203871.42992: entering _queue_task() for managed-node3/package_facts 13731 1727203871.43423: worker is 1 (out of 1 available) 13731 1727203871.43435: exiting _queue_task() for managed-node3/package_facts 13731 1727203871.43450: done queuing things up, now waiting for results queue to drain 13731 1727203871.43452: waiting for pending results... 13731 1727203871.43986: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which packages are installed 13731 1727203871.44023: in run() - task 028d2410-947f-82dc-c122-000000000b16 13731 1727203871.44046: variable 'ansible_search_path' from source: unknown 13731 1727203871.44055: variable 'ansible_search_path' from source: unknown 13731 1727203871.44105: calling self._execute() 13731 1727203871.44228: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203871.44241: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203871.44258: variable 'omit' from source: magic vars 13731 1727203871.44821: variable 'ansible_distribution_major_version' from source: facts 13731 1727203871.44879: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203871.44883: variable 'omit' from source: magic vars 13731 1727203871.44996: variable 'omit' from source: magic vars 13731 1727203871.45036: variable 'omit' from source: magic vars 13731 1727203871.45180: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203871.45184: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203871.45187: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203871.45190: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203871.45193: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203871.45226: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203871.45234: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203871.45241: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203871.45357: Set connection var ansible_pipelining to False 13731 1727203871.45369: Set connection var ansible_shell_type to sh 13731 1727203871.45383: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203871.45393: Set connection var ansible_connection to ssh 13731 1727203871.45403: Set connection var ansible_shell_executable to /bin/sh 13731 1727203871.45418: Set connection var ansible_timeout to 10 13731 1727203871.45447: variable 'ansible_shell_executable' from source: unknown 13731 1727203871.45457: variable 'ansible_connection' from source: unknown 13731 1727203871.45465: variable 'ansible_module_compression' from source: unknown 13731 1727203871.45472: variable 'ansible_shell_type' from source: unknown 13731 1727203871.45481: variable 'ansible_shell_executable' from source: unknown 13731 1727203871.45488: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203871.45495: variable 'ansible_pipelining' from source: unknown 13731 1727203871.45502: variable 'ansible_timeout' from source: unknown 13731 1727203871.45509: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203871.45725: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203871.45852: variable 'omit' from source: magic vars 13731 1727203871.45856: starting attempt loop 13731 1727203871.45859: running the handler 13731 1727203871.45861: _low_level_execute_command(): starting 13731 1727203871.45863: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203871.46761: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203871.46795: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203871.46812: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203871.46828: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203871.46913: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203871.48499: stdout chunk (state=3): >>>/root <<< 13731 1727203871.48885: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203871.48889: stdout chunk (state=3): >>><<< 13731 1727203871.48891: stderr chunk (state=3): >>><<< 13731 1727203871.48894: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203871.48896: _low_level_execute_command(): starting 13731 1727203871.48904: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122 `" && echo ansible-tmp-1727203871.4878085-17587-224361888383122="` echo /root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122 `" ) && sleep 0' 13731 1727203871.49795: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203871.49805: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203871.49834: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203871.49862: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203871.49899: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203871.51744: stdout chunk (state=3): >>>ansible-tmp-1727203871.4878085-17587-224361888383122=/root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122 <<< 13731 1727203871.51908: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203871.51912: stdout chunk (state=3): >>><<< 13731 1727203871.51914: stderr chunk (state=3): >>><<< 13731 1727203871.51940: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203871.4878085-17587-224361888383122=/root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203871.52047: variable 'ansible_module_compression' from source: unknown 13731 1727203871.52387: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.package_facts-ZIP_DEFLATED 13731 1727203871.52390: variable 'ansible_facts' from source: unknown 13731 1727203871.52721: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/AnsiballZ_package_facts.py 13731 1727203871.53066: Sending initial data 13731 1727203871.53080: Sent initial data (162 bytes) 13731 1727203871.53729: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203871.53792: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203871.53853: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203871.53873: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203871.53899: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203871.53969: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203871.55504: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203871.55581: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203871.55621: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmprjs782un /root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/AnsiballZ_package_facts.py <<< 13731 1727203871.55655: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/AnsiballZ_package_facts.py" <<< 13731 1727203871.55683: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmprjs782un" to remote "/root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/AnsiballZ_package_facts.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/AnsiballZ_package_facts.py" <<< 13731 1727203871.57328: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203871.57332: stdout chunk (state=3): >>><<< 13731 1727203871.57335: stderr chunk (state=3): >>><<< 13731 1727203871.57337: done transferring module to remote 13731 1727203871.57339: _low_level_execute_command(): starting 13731 1727203871.57342: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/ /root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/AnsiballZ_package_facts.py && sleep 0' 13731 1727203871.58017: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203871.58110: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203871.58149: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203871.58178: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203871.58192: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203871.58269: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203871.60149: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203871.60162: stdout chunk (state=3): >>><<< 13731 1727203871.60165: stderr chunk (state=3): >>><<< 13731 1727203871.60382: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203871.60386: _low_level_execute_command(): starting 13731 1727203871.60389: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/AnsiballZ_package_facts.py && sleep 0' 13731 1727203871.61380: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203871.61399: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203871.61447: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203871.61509: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203871.61539: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203871.61689: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203872.06035: stdout chunk (state=3): >>> {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "hunspell-filesystem": [{"name": "hunspell-filesystem", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "google-noto-fonts-common": [{"name": "google-noto-fonts-common", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-mono-vf-fonts": [{"name": "google-noto-sans-mono-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-vf-fonts": [{"name": "google-noto-sans-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-serif-vf-fonts": [{"name": "google-noto-serif-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-mono-vf-fonts": [{"name": "redhat-mono-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-text-vf-fonts": [{"name": "redhat-text-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "default-fonts-core-sans": [{"name": "default-fonts-core-sans", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-fonts-en": [{"name": "langpacks-fonts-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-ucode-firmware": [{"name": "amd-ucode-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "atheros-firmware": [{"name": "atheros-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "brcmfmac-firmware": [{"name": "brcmfmac-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "cirrus-audio-firmware": [{"name": "cirrus-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-audio-firmware": [{"name": "intel-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "mt7xxx-firmware": [{"name": "mt7xxx-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nxpwireless-firmware": [{"name": "nxpwireless-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "realtek-firmware": [{"name": "realtek-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tiwilink-firmware": [{"name": "tiwilink-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-gpu-firmware": [{"name": "amd-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-gpu-firmware": [{"name": "intel-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nvidia-gpu-firmware": [{"name": "nvidia-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "xkeyboard-config": [{"name": "xkeyboard-config", "version": "2.41", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-data": [{"name": "vim-data", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20240107", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.379", "release": "10.1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.14.5", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.18", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "21.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.2.26", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-ng-compat": [{"name": "zlib-ng-compat", "version": "2.1.6", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.19", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.5", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.5", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.36", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.30", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.50", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.11", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.46.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.4", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "1.1", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "51.20230828cvs.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.6.2", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.1", "release": "9.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.17", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.9", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.10.0", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.5.0", "release": "1.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "hunspell-en-GB": [{"name": "hunspell-en-GB", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell-en-US": [{"name": "hunspell-en-US", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell": [{"name": "hunspell", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.6.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam-libs": [{"name": "pam-libs", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.69", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.15.0", "release": "3.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.19.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.80.4", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-libs": [{"name": "polkit-libs", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "101.2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.13", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "31", "release": "8.el10", "epoch": null,<<< 13731 1727203872.06139: stdout chunk (state=3): >>> "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.11.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.6.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.7", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.10", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.3", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.6", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.4", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.7.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.16", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.6.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.5", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "42.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.7", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sequoia": [{"name": "rpm-sequoia", "version": "1.6.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.29", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.79.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsecret": [{"name": "libsecret", "version": "0.21.2", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pinentry": [{"name": "pinentry", "version": "1.3.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libusb1": [{"name": "libusb1", "version": "1.0.27", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "4.0.4", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hunspell-en": [{"name": "hunspell-en", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.2.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm": [{"name": "gdbm", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libxkbcommon": [{"name": "libxkbcommon", "version": "1.7.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.23.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.7", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.10", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.8", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "35", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "661", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.6", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "memstrack": [{"name": "memstrack", "version": "0.2.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.25.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.15", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "duktape": [{"name": "duktape", "version": "2.7.0", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "22.el10.gating_test1", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse3-libs": [{"name": "fuse3-libs", "version": "3.16.2", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-envsubst": [{"name": "gettext-envsubst", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-runtime": [{"name": "gettext-runtime", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "58", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.1.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.11.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.14.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "28.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.62.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.7", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-gssapi": [{"name": "cyrus-sasl-gssapi", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11^20190603git9e74f2d", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "polkit": [{"name": "polkit", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-pkla-compat": [{"name": "polkit-pkla-compat", "version": "0.1", "release": "29.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "audit-rules": [{"name": "audit-rules", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20240905", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hasher": [{"name": "libkcapi-hasher", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.22.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "makedumpfile": [{"name": "makedumpfile", "version": "1.7.5", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.9.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.18.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss-fapi": [{"name": "tpm2-tss-fapi", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tools": [{"name": "tpm2-tools", "version": "5.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "76.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.81", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.15.0", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.32", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-krb5-common": [{"name": "sssd-krb5-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpdecimal": [{"name": "mpdecimal", "version": "2.5.1", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.3.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo-python-plugin": [{"name": "sudo-python-plugin", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.16.0", "release": "15.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.2", "release": "14.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "235", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.9", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.7.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.29", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kdump-utils": [{"name": "kdump-utils", "version": "1.0.43", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.13.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-libs": [{"name": "pcsc-lite-libs", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-ccid": [<<< 13731 1727203872.06172: stdout chunk (state=3): >>>{"name": "pcsc-lite-ccid", "version": "1.6.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite": [{"name": "pcsc-lite", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2-smime": [{"name": "gnupg2-smime", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.3", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.24", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.14.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "1.2304.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2408.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "8.el10", "epoch": 18, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.12.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.2.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.20", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwlwifi-dvm-firmware": [{"name": "iwlwifi-dvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwlwifi-mvm-firmware": [{"name": "iwlwifi-mvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "37.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.5", "release": "0.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.20", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.59", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.89", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.189", "release": "511.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.15", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.27", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2024.001", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20231213", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.61.000", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-locale": [{"name": "perl-locale", "version": "1.12", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.42", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.350", "release": "510.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.085", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.68", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "2.20", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.22", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.45", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.088", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.038", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.09", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "5.01", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.18", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.31", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.29", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.55", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.03", "release": "510.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.38", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.86", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.63", "release": "510.el10", "epoch": 5, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.32", "release": "510.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.37", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.241", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.58", "release": "2.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.54", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.78", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.91", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.56", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.17", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.21", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.44", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "23.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch",<<< 13731 1727203872.06227: stdout chunk (state=3): >>> "source": "rpm"}], "git": [{"name": "git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xxd": [{"name": "xxd", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.39", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "5.2.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "24.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.35", "release": "4.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.10", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.24.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.4.1", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.9.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.3.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpds-py": [{"name": "python3-rpds-py", "version": "0.17.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "23.2.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-referencing": [{"name": "python3-referencing", "version": "0.31.1", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.19", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema-specifications": [{"name": "python3-jsonschema-specifications", "version": "2023.11.2", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "4.19.1", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.5", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.2.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "2.1.3", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "3.1.4", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.3", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.33", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.9.0", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.8", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "69.0.3", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.5.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "6.0.1", "release": "18.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-charset-normalizer": [{"name": "python3-charset-normalizer", "version": "3.3.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.32.3", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dhcpcd": [{"name": "dhcpcd", "version": "10.0.6", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "24.1.4", "release": "17.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "device-mapper-event-libs": [{"name": "device-mapper-event-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "libaio": [{"name": "libaio", "version": "0.3.111", "release": "20.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-event": [{"name": "device-mapper-event", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "lvm2-libs": [{"name": "lvm2-libs", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper-persistent-data": [{"name": "device-mapper-persistent-data", "version": "1.0.11", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lvm2": [{"name": "lvm2", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.17", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.90", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} <<< 13731 1727203872.08136: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203872.08140: stdout chunk (state=3): >>><<< 13731 1727203872.08143: stderr chunk (state=3): >>><<< 13731 1727203872.08389: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "hunspell-filesystem": [{"name": "hunspell-filesystem", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "google-noto-fonts-common": [{"name": "google-noto-fonts-common", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-mono-vf-fonts": [{"name": "google-noto-sans-mono-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-vf-fonts": [{"name": "google-noto-sans-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-serif-vf-fonts": [{"name": "google-noto-serif-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-mono-vf-fonts": [{"name": "redhat-mono-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-text-vf-fonts": [{"name": "redhat-text-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "default-fonts-core-sans": [{"name": "default-fonts-core-sans", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-fonts-en": [{"name": "langpacks-fonts-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-ucode-firmware": [{"name": "amd-ucode-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "atheros-firmware": [{"name": "atheros-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "brcmfmac-firmware": [{"name": "brcmfmac-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "cirrus-audio-firmware": [{"name": "cirrus-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-audio-firmware": [{"name": "intel-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "mt7xxx-firmware": [{"name": "mt7xxx-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nxpwireless-firmware": [{"name": "nxpwireless-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "realtek-firmware": [{"name": "realtek-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tiwilink-firmware": [{"name": "tiwilink-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-gpu-firmware": [{"name": "amd-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-gpu-firmware": [{"name": "intel-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nvidia-gpu-firmware": [{"name": "nvidia-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "xkeyboard-config": [{"name": "xkeyboard-config", "version": "2.41", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-data": [{"name": "vim-data", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20240107", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.379", "release": "10.1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.14.5", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.18", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "21.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.2.26", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-ng-compat": [{"name": "zlib-ng-compat", "version": "2.1.6", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.19", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.5", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.5", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.36", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.30", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.50", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.11", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.46.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.4", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "1.1", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "51.20230828cvs.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.6.2", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.1", "release": "9.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.17", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.9", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.10.0", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.5.0", "release": "1.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "hunspell-en-GB": [{"name": "hunspell-en-GB", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell-en-US": [{"name": "hunspell-en-US", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell": [{"name": "hunspell", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.6.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam-libs": [{"name": "pam-libs", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.69", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.15.0", "release": "3.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.19.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.80.4", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-libs": [{"name": "polkit-libs", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "101.2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.13", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.11.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.6.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.7", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.10", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.3", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.6", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.4", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.7.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.16", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.6.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.5", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "42.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.7", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sequoia": [{"name": "rpm-sequoia", "version": "1.6.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.29", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.79.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsecret": [{"name": "libsecret", "version": "0.21.2", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pinentry": [{"name": "pinentry", "version": "1.3.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libusb1": [{"name": "libusb1", "version": "1.0.27", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "4.0.4", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hunspell-en": [{"name": "hunspell-en", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.2.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm": [{"name": "gdbm", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libxkbcommon": [{"name": "libxkbcommon", "version": "1.7.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.23.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.7", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.10", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.8", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "35", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "661", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.6", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "memstrack": [{"name": "memstrack", "version": "0.2.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.25.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.15", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "duktape": [{"name": "duktape", "version": "2.7.0", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "22.el10.gating_test1", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse3-libs": [{"name": "fuse3-libs", "version": "3.16.2", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-envsubst": [{"name": "gettext-envsubst", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-runtime": [{"name": "gettext-runtime", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "58", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.1.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.11.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.14.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "28.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.62.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.7", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-gssapi": [{"name": "cyrus-sasl-gssapi", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11^20190603git9e74f2d", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "polkit": [{"name": "polkit", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-pkla-compat": [{"name": "polkit-pkla-compat", "version": "0.1", "release": "29.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "audit-rules": [{"name": "audit-rules", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20240905", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hasher": [{"name": "libkcapi-hasher", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.22.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "makedumpfile": [{"name": "makedumpfile", "version": "1.7.5", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.9.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.18.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss-fapi": [{"name": "tpm2-tss-fapi", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tools": [{"name": "tpm2-tools", "version": "5.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "76.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.81", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.15.0", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.32", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-krb5-common": [{"name": "sssd-krb5-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpdecimal": [{"name": "mpdecimal", "version": "2.5.1", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.3.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo-python-plugin": [{"name": "sudo-python-plugin", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.16.0", "release": "15.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.2", "release": "14.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "235", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.9", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.7.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.29", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kdump-utils": [{"name": "kdump-utils", "version": "1.0.43", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.13.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-libs": [{"name": "pcsc-lite-libs", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-ccid": [{"name": "pcsc-lite-ccid", "version": "1.6.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite": [{"name": "pcsc-lite", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2-smime": [{"name": "gnupg2-smime", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.3", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.24", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.14.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "1.2304.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2408.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "8.el10", "epoch": 18, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.12.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.2.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.20", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwlwifi-dvm-firmware": [{"name": "iwlwifi-dvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwlwifi-mvm-firmware": [{"name": "iwlwifi-mvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "37.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.5", "release": "0.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.20", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.59", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.89", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.189", "release": "511.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.15", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.27", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2024.001", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20231213", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.61.000", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-locale": [{"name": "perl-locale", "version": "1.12", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.42", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.350", "release": "510.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.085", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.68", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "2.20", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.22", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.45", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.088", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.038", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.09", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "5.01", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.18", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.31", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.29", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.55", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.03", "release": "510.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.38", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.86", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.63", "release": "510.el10", "epoch": 5, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.32", "release": "510.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.37", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.241", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.58", "release": "2.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.54", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.78", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.91", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.56", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.17", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.21", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.44", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "23.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xxd": [{"name": "xxd", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.39", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "5.2.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "24.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.35", "release": "4.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.10", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.24.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.4.1", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.9.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.3.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpds-py": [{"name": "python3-rpds-py", "version": "0.17.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "23.2.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-referencing": [{"name": "python3-referencing", "version": "0.31.1", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.19", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema-specifications": [{"name": "python3-jsonschema-specifications", "version": "2023.11.2", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "4.19.1", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.5", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.2.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "2.1.3", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "3.1.4", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.3", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.33", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.9.0", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.8", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "69.0.3", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.5.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "6.0.1", "release": "18.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-charset-normalizer": [{"name": "python3-charset-normalizer", "version": "3.3.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.32.3", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dhcpcd": [{"name": "dhcpcd", "version": "10.0.6", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "24.1.4", "release": "17.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "device-mapper-event-libs": [{"name": "device-mapper-event-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "libaio": [{"name": "libaio", "version": "0.3.111", "release": "20.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-event": [{"name": "device-mapper-event", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "lvm2-libs": [{"name": "lvm2-libs", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper-persistent-data": [{"name": "device-mapper-persistent-data", "version": "1.0.11", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lvm2": [{"name": "lvm2", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.17", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.90", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203872.23284: done with _execute_module (package_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'package_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203872.23343: _low_level_execute_command(): starting 13731 1727203872.23347: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203871.4878085-17587-224361888383122/ > /dev/null 2>&1 && sleep 0' 13731 1727203872.25084: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203872.25088: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203872.25107: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203872.25126: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203872.25186: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203872.25491: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203872.27382: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203872.27397: stdout chunk (state=3): >>><<< 13731 1727203872.27409: stderr chunk (state=3): >>><<< 13731 1727203872.27430: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203872.27442: handler run complete 13731 1727203872.28479: variable 'ansible_facts' from source: unknown 13731 1727203872.29011: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.31135: variable 'ansible_facts' from source: unknown 13731 1727203872.31899: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.33401: attempt loop complete, returning result 13731 1727203872.33417: _execute() done 13731 1727203872.33420: dumping result to json 13731 1727203872.33860: done dumping result, returning 13731 1727203872.33866: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which packages are installed [028d2410-947f-82dc-c122-000000000b16] 13731 1727203872.33869: sending task result for task 028d2410-947f-82dc-c122-000000000b16 13731 1727203872.47102: done sending task result for task 028d2410-947f-82dc-c122-000000000b16 13731 1727203872.47106: WORKER PROCESS EXITING ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203872.47178: no more pending results, returning what we have 13731 1727203872.47181: results queue empty 13731 1727203872.47182: checking for any_errors_fatal 13731 1727203872.47190: done checking for any_errors_fatal 13731 1727203872.47191: checking for max_fail_percentage 13731 1727203872.47192: done checking for max_fail_percentage 13731 1727203872.47193: checking to see if all hosts have failed and the running result is not ok 13731 1727203872.47194: done checking to see if all hosts have failed 13731 1727203872.47195: getting the remaining hosts for this loop 13731 1727203872.47196: done getting the remaining hosts for this loop 13731 1727203872.47199: getting the next task for host managed-node3 13731 1727203872.47205: done getting next task for host managed-node3 13731 1727203872.47208: ^ task is: TASK: fedora.linux_system_roles.network : Print network provider 13731 1727203872.47214: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203872.47224: getting variables 13731 1727203872.47226: in VariableManager get_vars() 13731 1727203872.47249: Calling all_inventory to load vars for managed-node3 13731 1727203872.47252: Calling groups_inventory to load vars for managed-node3 13731 1727203872.47254: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203872.47260: Calling all_plugins_play to load vars for managed-node3 13731 1727203872.47263: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203872.47266: Calling groups_plugins_play to load vars for managed-node3 13731 1727203872.49043: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.50726: done with get_vars() 13731 1727203872.50750: done getting variables 13731 1727203872.50808: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Print network provider] ************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:7 Tuesday 24 September 2024 14:51:12 -0400 (0:00:01.078) 0:00:50.739 ***** 13731 1727203872.50847: entering _queue_task() for managed-node3/debug 13731 1727203872.51211: worker is 1 (out of 1 available) 13731 1727203872.51224: exiting _queue_task() for managed-node3/debug 13731 1727203872.51235: done queuing things up, now waiting for results queue to drain 13731 1727203872.51237: waiting for pending results... 13731 1727203872.51604: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Print network provider 13731 1727203872.51710: in run() - task 028d2410-947f-82dc-c122-000000000a2f 13731 1727203872.51734: variable 'ansible_search_path' from source: unknown 13731 1727203872.51742: variable 'ansible_search_path' from source: unknown 13731 1727203872.51787: calling self._execute() 13731 1727203872.51917: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.51922: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.51935: variable 'omit' from source: magic vars 13731 1727203872.52464: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.52467: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203872.52470: variable 'omit' from source: magic vars 13731 1727203872.52472: variable 'omit' from source: magic vars 13731 1727203872.52560: variable 'network_provider' from source: set_fact 13731 1727203872.52589: variable 'omit' from source: magic vars 13731 1727203872.52635: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203872.52678: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203872.52706: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203872.52728: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203872.52746: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203872.52785: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203872.52798: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.52806: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.52916: Set connection var ansible_pipelining to False 13731 1727203872.52926: Set connection var ansible_shell_type to sh 13731 1727203872.52935: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203872.52943: Set connection var ansible_connection to ssh 13731 1727203872.52953: Set connection var ansible_shell_executable to /bin/sh 13731 1727203872.52981: Set connection var ansible_timeout to 10 13731 1727203872.52997: variable 'ansible_shell_executable' from source: unknown 13731 1727203872.53005: variable 'ansible_connection' from source: unknown 13731 1727203872.53080: variable 'ansible_module_compression' from source: unknown 13731 1727203872.53084: variable 'ansible_shell_type' from source: unknown 13731 1727203872.53086: variable 'ansible_shell_executable' from source: unknown 13731 1727203872.53088: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.53090: variable 'ansible_pipelining' from source: unknown 13731 1727203872.53092: variable 'ansible_timeout' from source: unknown 13731 1727203872.53094: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.53203: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203872.53219: variable 'omit' from source: magic vars 13731 1727203872.53232: starting attempt loop 13731 1727203872.53240: running the handler 13731 1727203872.53294: handler run complete 13731 1727203872.53312: attempt loop complete, returning result 13731 1727203872.53319: _execute() done 13731 1727203872.53324: dumping result to json 13731 1727203872.53331: done dumping result, returning 13731 1727203872.53346: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Print network provider [028d2410-947f-82dc-c122-000000000a2f] 13731 1727203872.53355: sending task result for task 028d2410-947f-82dc-c122-000000000a2f 13731 1727203872.53520: done sending task result for task 028d2410-947f-82dc-c122-000000000a2f 13731 1727203872.53524: WORKER PROCESS EXITING ok: [managed-node3] => {} MSG: Using network provider: nm 13731 1727203872.53623: no more pending results, returning what we have 13731 1727203872.53628: results queue empty 13731 1727203872.53629: checking for any_errors_fatal 13731 1727203872.53642: done checking for any_errors_fatal 13731 1727203872.53643: checking for max_fail_percentage 13731 1727203872.53645: done checking for max_fail_percentage 13731 1727203872.53646: checking to see if all hosts have failed and the running result is not ok 13731 1727203872.53647: done checking to see if all hosts have failed 13731 1727203872.53647: getting the remaining hosts for this loop 13731 1727203872.53649: done getting the remaining hosts for this loop 13731 1727203872.53653: getting the next task for host managed-node3 13731 1727203872.53665: done getting next task for host managed-node3 13731 1727203872.53668: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13731 1727203872.53674: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203872.53687: getting variables 13731 1727203872.53689: in VariableManager get_vars() 13731 1727203872.53734: Calling all_inventory to load vars for managed-node3 13731 1727203872.53737: Calling groups_inventory to load vars for managed-node3 13731 1727203872.53740: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203872.53751: Calling all_plugins_play to load vars for managed-node3 13731 1727203872.53753: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203872.53757: Calling groups_plugins_play to load vars for managed-node3 13731 1727203872.55721: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.57877: done with get_vars() 13731 1727203872.57924: done getting variables 13731 1727203872.58014: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:11 Tuesday 24 September 2024 14:51:12 -0400 (0:00:00.072) 0:00:50.811 ***** 13731 1727203872.58082: entering _queue_task() for managed-node3/fail 13731 1727203872.58853: worker is 1 (out of 1 available) 13731 1727203872.58868: exiting _queue_task() for managed-node3/fail 13731 1727203872.58881: done queuing things up, now waiting for results queue to drain 13731 1727203872.58883: waiting for pending results... 13731 1727203872.59183: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13731 1727203872.59387: in run() - task 028d2410-947f-82dc-c122-000000000a30 13731 1727203872.59408: variable 'ansible_search_path' from source: unknown 13731 1727203872.59414: variable 'ansible_search_path' from source: unknown 13731 1727203872.59456: calling self._execute() 13731 1727203872.59591: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.59603: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.59614: variable 'omit' from source: magic vars 13731 1727203872.60040: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.60053: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203872.60190: variable 'network_state' from source: role '' defaults 13731 1727203872.60242: Evaluated conditional (network_state != {}): False 13731 1727203872.60246: when evaluation is False, skipping this task 13731 1727203872.60248: _execute() done 13731 1727203872.60252: dumping result to json 13731 1727203872.60254: done dumping result, returning 13731 1727203872.60257: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider [028d2410-947f-82dc-c122-000000000a30] 13731 1727203872.60261: sending task result for task 028d2410-947f-82dc-c122-000000000a30 skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203872.60417: no more pending results, returning what we have 13731 1727203872.60421: results queue empty 13731 1727203872.60422: checking for any_errors_fatal 13731 1727203872.60430: done checking for any_errors_fatal 13731 1727203872.60430: checking for max_fail_percentage 13731 1727203872.60432: done checking for max_fail_percentage 13731 1727203872.60433: checking to see if all hosts have failed and the running result is not ok 13731 1727203872.60434: done checking to see if all hosts have failed 13731 1727203872.60434: getting the remaining hosts for this loop 13731 1727203872.60436: done getting the remaining hosts for this loop 13731 1727203872.60440: getting the next task for host managed-node3 13731 1727203872.60449: done getting next task for host managed-node3 13731 1727203872.60453: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13731 1727203872.60461: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203872.60487: getting variables 13731 1727203872.60489: in VariableManager get_vars() 13731 1727203872.60534: Calling all_inventory to load vars for managed-node3 13731 1727203872.60537: Calling groups_inventory to load vars for managed-node3 13731 1727203872.60539: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203872.60551: Calling all_plugins_play to load vars for managed-node3 13731 1727203872.60554: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203872.60557: Calling groups_plugins_play to load vars for managed-node3 13731 1727203872.61493: done sending task result for task 028d2410-947f-82dc-c122-000000000a30 13731 1727203872.61497: WORKER PROCESS EXITING 13731 1727203872.62473: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.64701: done with get_vars() 13731 1727203872.64736: done getting variables 13731 1727203872.64799: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:18 Tuesday 24 September 2024 14:51:12 -0400 (0:00:00.067) 0:00:50.879 ***** 13731 1727203872.64836: entering _queue_task() for managed-node3/fail 13731 1727203872.65469: worker is 1 (out of 1 available) 13731 1727203872.65492: exiting _queue_task() for managed-node3/fail 13731 1727203872.65505: done queuing things up, now waiting for results queue to drain 13731 1727203872.65506: waiting for pending results... 13731 1727203872.65693: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13731 1727203872.65807: in run() - task 028d2410-947f-82dc-c122-000000000a31 13731 1727203872.65819: variable 'ansible_search_path' from source: unknown 13731 1727203872.65822: variable 'ansible_search_path' from source: unknown 13731 1727203872.65853: calling self._execute() 13731 1727203872.65940: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.65946: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.65956: variable 'omit' from source: magic vars 13731 1727203872.66239: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.66249: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203872.66334: variable 'network_state' from source: role '' defaults 13731 1727203872.66343: Evaluated conditional (network_state != {}): False 13731 1727203872.66347: when evaluation is False, skipping this task 13731 1727203872.66349: _execute() done 13731 1727203872.66352: dumping result to json 13731 1727203872.66354: done dumping result, returning 13731 1727203872.66364: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 [028d2410-947f-82dc-c122-000000000a31] 13731 1727203872.66367: sending task result for task 028d2410-947f-82dc-c122-000000000a31 13731 1727203872.66466: done sending task result for task 028d2410-947f-82dc-c122-000000000a31 13731 1727203872.66469: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203872.66543: no more pending results, returning what we have 13731 1727203872.66547: results queue empty 13731 1727203872.66548: checking for any_errors_fatal 13731 1727203872.66554: done checking for any_errors_fatal 13731 1727203872.66555: checking for max_fail_percentage 13731 1727203872.66557: done checking for max_fail_percentage 13731 1727203872.66560: checking to see if all hosts have failed and the running result is not ok 13731 1727203872.66561: done checking to see if all hosts have failed 13731 1727203872.66562: getting the remaining hosts for this loop 13731 1727203872.66564: done getting the remaining hosts for this loop 13731 1727203872.66567: getting the next task for host managed-node3 13731 1727203872.66577: done getting next task for host managed-node3 13731 1727203872.66581: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13731 1727203872.66585: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203872.66606: getting variables 13731 1727203872.66608: in VariableManager get_vars() 13731 1727203872.66643: Calling all_inventory to load vars for managed-node3 13731 1727203872.66646: Calling groups_inventory to load vars for managed-node3 13731 1727203872.66648: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203872.66657: Calling all_plugins_play to load vars for managed-node3 13731 1727203872.66662: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203872.66665: Calling groups_plugins_play to load vars for managed-node3 13731 1727203872.67755: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.68883: done with get_vars() 13731 1727203872.68908: done getting variables 13731 1727203872.68952: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:25 Tuesday 24 September 2024 14:51:12 -0400 (0:00:00.041) 0:00:50.920 ***** 13731 1727203872.68982: entering _queue_task() for managed-node3/fail 13731 1727203872.69249: worker is 1 (out of 1 available) 13731 1727203872.69264: exiting _queue_task() for managed-node3/fail 13731 1727203872.69278: done queuing things up, now waiting for results queue to drain 13731 1727203872.69280: waiting for pending results... 13731 1727203872.69473: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13731 1727203872.69593: in run() - task 028d2410-947f-82dc-c122-000000000a32 13731 1727203872.69606: variable 'ansible_search_path' from source: unknown 13731 1727203872.69612: variable 'ansible_search_path' from source: unknown 13731 1727203872.69642: calling self._execute() 13731 1727203872.69732: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.69736: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.69739: variable 'omit' from source: magic vars 13731 1727203872.70036: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.70045: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203872.70184: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203872.72858: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203872.72966: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203872.72981: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203872.73023: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203872.73056: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203872.73182: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.73191: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.73221: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.73266: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.73290: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.73399: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.73581: Evaluated conditional (ansible_distribution_major_version | int > 9): True 13731 1727203872.73584: variable 'ansible_distribution' from source: facts 13731 1727203872.73587: variable '__network_rh_distros' from source: role '' defaults 13731 1727203872.73589: Evaluated conditional (ansible_distribution in __network_rh_distros): True 13731 1727203872.73832: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.73892: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.73925: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.73969: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.73990: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.74046: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.74074: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.74104: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.74155: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.74177: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.74224: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.74259: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.74291: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.74333: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.74356: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.74688: variable 'network_connections' from source: task vars 13731 1727203872.74707: variable 'controller_profile' from source: play vars 13731 1727203872.74786: variable 'controller_profile' from source: play vars 13731 1727203872.74806: variable 'controller_device' from source: play vars 13731 1727203872.74902: variable 'controller_device' from source: play vars 13731 1727203872.74906: variable 'dhcp_interface1' from source: play vars 13731 1727203872.74949: variable 'dhcp_interface1' from source: play vars 13731 1727203872.74963: variable 'port1_profile' from source: play vars 13731 1727203872.75028: variable 'port1_profile' from source: play vars 13731 1727203872.75040: variable 'dhcp_interface1' from source: play vars 13731 1727203872.75120: variable 'dhcp_interface1' from source: play vars 13731 1727203872.75124: variable 'controller_profile' from source: play vars 13731 1727203872.75178: variable 'controller_profile' from source: play vars 13731 1727203872.75229: variable 'port2_profile' from source: play vars 13731 1727203872.75254: variable 'port2_profile' from source: play vars 13731 1727203872.75267: variable 'dhcp_interface2' from source: play vars 13731 1727203872.75329: variable 'dhcp_interface2' from source: play vars 13731 1727203872.75356: variable 'controller_profile' from source: play vars 13731 1727203872.75453: variable 'controller_profile' from source: play vars 13731 1727203872.75456: variable 'network_state' from source: role '' defaults 13731 1727203872.75505: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203872.75691: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203872.75735: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203872.75774: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203872.75980: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203872.75984: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203872.75986: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203872.75988: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.75990: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203872.75993: Evaluated conditional (network_connections | selectattr("type", "defined") | selectattr("type", "match", "^team$") | list | length > 0 or network_state.get("interfaces", []) | selectattr("type", "defined") | selectattr("type", "match", "^team$") | list | length > 0): False 13731 1727203872.75995: when evaluation is False, skipping this task 13731 1727203872.75997: _execute() done 13731 1727203872.75999: dumping result to json 13731 1727203872.76001: done dumping result, returning 13731 1727203872.76003: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later [028d2410-947f-82dc-c122-000000000a32] 13731 1727203872.76012: sending task result for task 028d2410-947f-82dc-c122-000000000a32 skipping: [managed-node3] => { "changed": false, "false_condition": "network_connections | selectattr(\"type\", \"defined\") | selectattr(\"type\", \"match\", \"^team$\") | list | length > 0 or network_state.get(\"interfaces\", []) | selectattr(\"type\", \"defined\") | selectattr(\"type\", \"match\", \"^team$\") | list | length > 0", "skip_reason": "Conditional result was False" } 13731 1727203872.76165: no more pending results, returning what we have 13731 1727203872.76170: results queue empty 13731 1727203872.76171: checking for any_errors_fatal 13731 1727203872.76179: done checking for any_errors_fatal 13731 1727203872.76180: checking for max_fail_percentage 13731 1727203872.76182: done checking for max_fail_percentage 13731 1727203872.76183: checking to see if all hosts have failed and the running result is not ok 13731 1727203872.76184: done checking to see if all hosts have failed 13731 1727203872.76185: getting the remaining hosts for this loop 13731 1727203872.76187: done getting the remaining hosts for this loop 13731 1727203872.76191: getting the next task for host managed-node3 13731 1727203872.76200: done getting next task for host managed-node3 13731 1727203872.76205: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13731 1727203872.76209: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203872.76230: getting variables 13731 1727203872.76232: in VariableManager get_vars() 13731 1727203872.76486: Calling all_inventory to load vars for managed-node3 13731 1727203872.76490: Calling groups_inventory to load vars for managed-node3 13731 1727203872.76493: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203872.76504: Calling all_plugins_play to load vars for managed-node3 13731 1727203872.76507: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203872.76510: Calling groups_plugins_play to load vars for managed-node3 13731 1727203872.77191: done sending task result for task 028d2410-947f-82dc-c122-000000000a32 13731 1727203872.77195: WORKER PROCESS EXITING 13731 1727203872.78188: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.79724: done with get_vars() 13731 1727203872.79757: done getting variables 13731 1727203872.79819: Loading ActionModule 'dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:36 Tuesday 24 September 2024 14:51:12 -0400 (0:00:00.108) 0:00:51.029 ***** 13731 1727203872.79857: entering _queue_task() for managed-node3/dnf 13731 1727203872.80216: worker is 1 (out of 1 available) 13731 1727203872.80229: exiting _queue_task() for managed-node3/dnf 13731 1727203872.80242: done queuing things up, now waiting for results queue to drain 13731 1727203872.80243: waiting for pending results... 13731 1727203872.80545: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13731 1727203872.80717: in run() - task 028d2410-947f-82dc-c122-000000000a33 13731 1727203872.80738: variable 'ansible_search_path' from source: unknown 13731 1727203872.80744: variable 'ansible_search_path' from source: unknown 13731 1727203872.80787: calling self._execute() 13731 1727203872.80890: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.80903: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.80921: variable 'omit' from source: magic vars 13731 1727203872.81320: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.81338: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203872.81544: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203872.83839: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203872.83922: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203872.83969: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203872.84012: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203872.84043: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203872.84130: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.84184: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.84217: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.84263: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.84288: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.84415: variable 'ansible_distribution' from source: facts 13731 1727203872.84424: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.84445: Evaluated conditional (ansible_distribution == 'Fedora' or ansible_distribution_major_version | int > 7): True 13731 1727203872.84571: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203872.84711: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.84744: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.84773: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.84820: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.84949: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.84952: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.84954: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.84957: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.84981: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.85000: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.85041: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.85077: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.85107: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.85149: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.85174: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.85328: variable 'network_connections' from source: task vars 13731 1727203872.85338: variable 'controller_profile' from source: play vars 13731 1727203872.85390: variable 'controller_profile' from source: play vars 13731 1727203872.85399: variable 'controller_device' from source: play vars 13731 1727203872.85441: variable 'controller_device' from source: play vars 13731 1727203872.85449: variable 'dhcp_interface1' from source: play vars 13731 1727203872.85506: variable 'dhcp_interface1' from source: play vars 13731 1727203872.85513: variable 'port1_profile' from source: play vars 13731 1727203872.85554: variable 'port1_profile' from source: play vars 13731 1727203872.85563: variable 'dhcp_interface1' from source: play vars 13731 1727203872.85604: variable 'dhcp_interface1' from source: play vars 13731 1727203872.85613: variable 'controller_profile' from source: play vars 13731 1727203872.85653: variable 'controller_profile' from source: play vars 13731 1727203872.85661: variable 'port2_profile' from source: play vars 13731 1727203872.85702: variable 'port2_profile' from source: play vars 13731 1727203872.85706: variable 'dhcp_interface2' from source: play vars 13731 1727203872.85750: variable 'dhcp_interface2' from source: play vars 13731 1727203872.85765: variable 'controller_profile' from source: play vars 13731 1727203872.85802: variable 'controller_profile' from source: play vars 13731 1727203872.85855: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203872.85982: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203872.86010: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203872.86034: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203872.86060: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203872.86104: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203872.86121: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203872.86138: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.86162: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203872.86208: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203872.86357: variable 'network_connections' from source: task vars 13731 1727203872.86365: variable 'controller_profile' from source: play vars 13731 1727203872.86409: variable 'controller_profile' from source: play vars 13731 1727203872.86415: variable 'controller_device' from source: play vars 13731 1727203872.86456: variable 'controller_device' from source: play vars 13731 1727203872.86463: variable 'dhcp_interface1' from source: play vars 13731 1727203872.86508: variable 'dhcp_interface1' from source: play vars 13731 1727203872.86515: variable 'port1_profile' from source: play vars 13731 1727203872.86555: variable 'port1_profile' from source: play vars 13731 1727203872.86563: variable 'dhcp_interface1' from source: play vars 13731 1727203872.86608: variable 'dhcp_interface1' from source: play vars 13731 1727203872.86613: variable 'controller_profile' from source: play vars 13731 1727203872.86654: variable 'controller_profile' from source: play vars 13731 1727203872.86663: variable 'port2_profile' from source: play vars 13731 1727203872.86706: variable 'port2_profile' from source: play vars 13731 1727203872.86711: variable 'dhcp_interface2' from source: play vars 13731 1727203872.86753: variable 'dhcp_interface2' from source: play vars 13731 1727203872.86761: variable 'controller_profile' from source: play vars 13731 1727203872.86802: variable 'controller_profile' from source: play vars 13731 1727203872.86830: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203872.86833: when evaluation is False, skipping this task 13731 1727203872.86836: _execute() done 13731 1727203872.86838: dumping result to json 13731 1727203872.86840: done dumping result, returning 13731 1727203872.86848: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000a33] 13731 1727203872.86852: sending task result for task 028d2410-947f-82dc-c122-000000000a33 13731 1727203872.86946: done sending task result for task 028d2410-947f-82dc-c122-000000000a33 13731 1727203872.86949: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203872.87010: no more pending results, returning what we have 13731 1727203872.87014: results queue empty 13731 1727203872.87015: checking for any_errors_fatal 13731 1727203872.87025: done checking for any_errors_fatal 13731 1727203872.87026: checking for max_fail_percentage 13731 1727203872.87028: done checking for max_fail_percentage 13731 1727203872.87029: checking to see if all hosts have failed and the running result is not ok 13731 1727203872.87030: done checking to see if all hosts have failed 13731 1727203872.87031: getting the remaining hosts for this loop 13731 1727203872.87033: done getting the remaining hosts for this loop 13731 1727203872.87036: getting the next task for host managed-node3 13731 1727203872.87101: done getting next task for host managed-node3 13731 1727203872.87106: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13731 1727203872.87111: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203872.87134: getting variables 13731 1727203872.87136: in VariableManager get_vars() 13731 1727203872.87219: Calling all_inventory to load vars for managed-node3 13731 1727203872.87222: Calling groups_inventory to load vars for managed-node3 13731 1727203872.87224: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203872.87233: Calling all_plugins_play to load vars for managed-node3 13731 1727203872.87236: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203872.87238: Calling groups_plugins_play to load vars for managed-node3 13731 1727203872.88372: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.90084: done with get_vars() 13731 1727203872.90119: done getting variables redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13731 1727203872.90220: Loading ActionModule 'ansible_collections.ansible.builtin.plugins.action.dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:48 Tuesday 24 September 2024 14:51:12 -0400 (0:00:00.103) 0:00:51.133 ***** 13731 1727203872.90262: entering _queue_task() for managed-node3/yum 13731 1727203872.90682: worker is 1 (out of 1 available) 13731 1727203872.90695: exiting _queue_task() for managed-node3/yum 13731 1727203872.90709: done queuing things up, now waiting for results queue to drain 13731 1727203872.90718: waiting for pending results... 13731 1727203872.91179: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13731 1727203872.91304: in run() - task 028d2410-947f-82dc-c122-000000000a34 13731 1727203872.91627: variable 'ansible_search_path' from source: unknown 13731 1727203872.91631: variable 'ansible_search_path' from source: unknown 13731 1727203872.91633: calling self._execute() 13731 1727203872.91747: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.91763: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.91782: variable 'omit' from source: magic vars 13731 1727203872.92230: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.92262: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203872.92457: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203872.95253: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203872.95302: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203872.95330: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203872.95355: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203872.95378: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203872.95442: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203872.95467: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203872.95486: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203872.95512: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203872.95528: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203872.95596: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.95608: Evaluated conditional (ansible_distribution_major_version | int < 8): False 13731 1727203872.95611: when evaluation is False, skipping this task 13731 1727203872.95614: _execute() done 13731 1727203872.95616: dumping result to json 13731 1727203872.95619: done dumping result, returning 13731 1727203872.95628: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000a34] 13731 1727203872.95632: sending task result for task 028d2410-947f-82dc-c122-000000000a34 13731 1727203872.95724: done sending task result for task 028d2410-947f-82dc-c122-000000000a34 13731 1727203872.95727: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version | int < 8", "skip_reason": "Conditional result was False" } 13731 1727203872.95793: no more pending results, returning what we have 13731 1727203872.95797: results queue empty 13731 1727203872.95798: checking for any_errors_fatal 13731 1727203872.95804: done checking for any_errors_fatal 13731 1727203872.95804: checking for max_fail_percentage 13731 1727203872.95806: done checking for max_fail_percentage 13731 1727203872.95807: checking to see if all hosts have failed and the running result is not ok 13731 1727203872.95808: done checking to see if all hosts have failed 13731 1727203872.95808: getting the remaining hosts for this loop 13731 1727203872.95810: done getting the remaining hosts for this loop 13731 1727203872.95813: getting the next task for host managed-node3 13731 1727203872.95822: done getting next task for host managed-node3 13731 1727203872.95826: ^ task is: TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13731 1727203872.95831: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203872.95849: getting variables 13731 1727203872.95852: in VariableManager get_vars() 13731 1727203872.95895: Calling all_inventory to load vars for managed-node3 13731 1727203872.95898: Calling groups_inventory to load vars for managed-node3 13731 1727203872.95900: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203872.95908: Calling all_plugins_play to load vars for managed-node3 13731 1727203872.95911: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203872.95913: Calling groups_plugins_play to load vars for managed-node3 13731 1727203872.96836: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203872.97717: done with get_vars() 13731 1727203872.97738: done getting variables 13731 1727203872.97786: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:60 Tuesday 24 September 2024 14:51:12 -0400 (0:00:00.075) 0:00:51.209 ***** 13731 1727203872.97812: entering _queue_task() for managed-node3/fail 13731 1727203872.98074: worker is 1 (out of 1 available) 13731 1727203872.98091: exiting _queue_task() for managed-node3/fail 13731 1727203872.98103: done queuing things up, now waiting for results queue to drain 13731 1727203872.98104: waiting for pending results... 13731 1727203872.98288: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13731 1727203872.98403: in run() - task 028d2410-947f-82dc-c122-000000000a35 13731 1727203872.98415: variable 'ansible_search_path' from source: unknown 13731 1727203872.98418: variable 'ansible_search_path' from source: unknown 13731 1727203872.98450: calling self._execute() 13731 1727203872.98523: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203872.98528: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203872.98535: variable 'omit' from source: magic vars 13731 1727203872.98810: variable 'ansible_distribution_major_version' from source: facts 13731 1727203872.98819: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203872.98905: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203872.99036: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203873.00515: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203873.00564: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203873.00591: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203873.00619: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203873.00639: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203873.00701: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.00738: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.00755: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.00784: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.00795: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.00829: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.00847: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.00864: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.00890: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.00901: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.00929: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.00947: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.00965: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.00991: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.01001: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.01122: variable 'network_connections' from source: task vars 13731 1727203873.01131: variable 'controller_profile' from source: play vars 13731 1727203873.01184: variable 'controller_profile' from source: play vars 13731 1727203873.01193: variable 'controller_device' from source: play vars 13731 1727203873.01237: variable 'controller_device' from source: play vars 13731 1727203873.01245: variable 'dhcp_interface1' from source: play vars 13731 1727203873.01290: variable 'dhcp_interface1' from source: play vars 13731 1727203873.01297: variable 'port1_profile' from source: play vars 13731 1727203873.01338: variable 'port1_profile' from source: play vars 13731 1727203873.01344: variable 'dhcp_interface1' from source: play vars 13731 1727203873.01387: variable 'dhcp_interface1' from source: play vars 13731 1727203873.01393: variable 'controller_profile' from source: play vars 13731 1727203873.01434: variable 'controller_profile' from source: play vars 13731 1727203873.01440: variable 'port2_profile' from source: play vars 13731 1727203873.01486: variable 'port2_profile' from source: play vars 13731 1727203873.01496: variable 'dhcp_interface2' from source: play vars 13731 1727203873.01534: variable 'dhcp_interface2' from source: play vars 13731 1727203873.01541: variable 'controller_profile' from source: play vars 13731 1727203873.01583: variable 'controller_profile' from source: play vars 13731 1727203873.01633: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203873.01744: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203873.01772: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203873.01796: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203873.01820: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203873.01852: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203873.01869: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203873.01888: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.01909: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203873.01957: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203873.02119: variable 'network_connections' from source: task vars 13731 1727203873.02122: variable 'controller_profile' from source: play vars 13731 1727203873.02168: variable 'controller_profile' from source: play vars 13731 1727203873.02174: variable 'controller_device' from source: play vars 13731 1727203873.02216: variable 'controller_device' from source: play vars 13731 1727203873.02223: variable 'dhcp_interface1' from source: play vars 13731 1727203873.02267: variable 'dhcp_interface1' from source: play vars 13731 1727203873.02274: variable 'port1_profile' from source: play vars 13731 1727203873.02316: variable 'port1_profile' from source: play vars 13731 1727203873.02322: variable 'dhcp_interface1' from source: play vars 13731 1727203873.02369: variable 'dhcp_interface1' from source: play vars 13731 1727203873.02372: variable 'controller_profile' from source: play vars 13731 1727203873.02411: variable 'controller_profile' from source: play vars 13731 1727203873.02417: variable 'port2_profile' from source: play vars 13731 1727203873.02457: variable 'port2_profile' from source: play vars 13731 1727203873.02463: variable 'dhcp_interface2' from source: play vars 13731 1727203873.02508: variable 'dhcp_interface2' from source: play vars 13731 1727203873.02512: variable 'controller_profile' from source: play vars 13731 1727203873.02552: variable 'controller_profile' from source: play vars 13731 1727203873.02583: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203873.02586: when evaluation is False, skipping this task 13731 1727203873.02589: _execute() done 13731 1727203873.02591: dumping result to json 13731 1727203873.02593: done dumping result, returning 13731 1727203873.02595: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000a35] 13731 1727203873.02600: sending task result for task 028d2410-947f-82dc-c122-000000000a35 13731 1727203873.02691: done sending task result for task 028d2410-947f-82dc-c122-000000000a35 13731 1727203873.02693: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203873.02747: no more pending results, returning what we have 13731 1727203873.02751: results queue empty 13731 1727203873.02752: checking for any_errors_fatal 13731 1727203873.02761: done checking for any_errors_fatal 13731 1727203873.02762: checking for max_fail_percentage 13731 1727203873.02764: done checking for max_fail_percentage 13731 1727203873.02765: checking to see if all hosts have failed and the running result is not ok 13731 1727203873.02765: done checking to see if all hosts have failed 13731 1727203873.02766: getting the remaining hosts for this loop 13731 1727203873.02768: done getting the remaining hosts for this loop 13731 1727203873.02771: getting the next task for host managed-node3 13731 1727203873.02781: done getting next task for host managed-node3 13731 1727203873.02784: ^ task is: TASK: fedora.linux_system_roles.network : Install packages 13731 1727203873.02789: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203873.02808: getting variables 13731 1727203873.02810: in VariableManager get_vars() 13731 1727203873.02861: Calling all_inventory to load vars for managed-node3 13731 1727203873.02864: Calling groups_inventory to load vars for managed-node3 13731 1727203873.02866: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203873.02874: Calling all_plugins_play to load vars for managed-node3 13731 1727203873.02879: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203873.02881: Calling groups_plugins_play to load vars for managed-node3 13731 1727203873.03694: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203873.04673: done with get_vars() 13731 1727203873.04692: done getting variables 13731 1727203873.04735: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install packages] ******************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:73 Tuesday 24 September 2024 14:51:13 -0400 (0:00:00.069) 0:00:51.278 ***** 13731 1727203873.04762: entering _queue_task() for managed-node3/package 13731 1727203873.05020: worker is 1 (out of 1 available) 13731 1727203873.05034: exiting _queue_task() for managed-node3/package 13731 1727203873.05047: done queuing things up, now waiting for results queue to drain 13731 1727203873.05048: waiting for pending results... 13731 1727203873.05241: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install packages 13731 1727203873.05349: in run() - task 028d2410-947f-82dc-c122-000000000a36 13731 1727203873.05360: variable 'ansible_search_path' from source: unknown 13731 1727203873.05364: variable 'ansible_search_path' from source: unknown 13731 1727203873.05399: calling self._execute() 13731 1727203873.05474: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203873.05479: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203873.05489: variable 'omit' from source: magic vars 13731 1727203873.05767: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.05778: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203873.05915: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203873.06112: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203873.06150: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203873.06176: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203873.06229: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203873.06314: variable 'network_packages' from source: role '' defaults 13731 1727203873.06389: variable '__network_provider_setup' from source: role '' defaults 13731 1727203873.06398: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203873.06441: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203873.06448: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203873.06498: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203873.06613: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203873.07954: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203873.08003: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203873.08030: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203873.08053: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203873.08078: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203873.08137: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.08157: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.08179: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.08209: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.08220: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.08252: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.08270: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.08288: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.08317: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.08327: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.08481: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13731 1727203873.08554: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.08574: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.08592: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.08616: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.08628: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.08691: variable 'ansible_python' from source: facts 13731 1727203873.08704: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13731 1727203873.08760: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203873.08817: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203873.08903: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.08919: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.08936: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.08960: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.08976: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.09008: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.09027: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.09043: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.09072: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.09084: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.09177: variable 'network_connections' from source: task vars 13731 1727203873.09181: variable 'controller_profile' from source: play vars 13731 1727203873.09250: variable 'controller_profile' from source: play vars 13731 1727203873.09259: variable 'controller_device' from source: play vars 13731 1727203873.09331: variable 'controller_device' from source: play vars 13731 1727203873.09340: variable 'dhcp_interface1' from source: play vars 13731 1727203873.09409: variable 'dhcp_interface1' from source: play vars 13731 1727203873.09412: variable 'port1_profile' from source: play vars 13731 1727203873.09504: variable 'port1_profile' from source: play vars 13731 1727203873.09512: variable 'dhcp_interface1' from source: play vars 13731 1727203873.09584: variable 'dhcp_interface1' from source: play vars 13731 1727203873.09592: variable 'controller_profile' from source: play vars 13731 1727203873.09660: variable 'controller_profile' from source: play vars 13731 1727203873.09670: variable 'port2_profile' from source: play vars 13731 1727203873.09741: variable 'port2_profile' from source: play vars 13731 1727203873.09744: variable 'dhcp_interface2' from source: play vars 13731 1727203873.09815: variable 'dhcp_interface2' from source: play vars 13731 1727203873.09822: variable 'controller_profile' from source: play vars 13731 1727203873.09897: variable 'controller_profile' from source: play vars 13731 1727203873.09952: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203873.09978: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203873.09998: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.10019: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203873.10057: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203873.10240: variable 'network_connections' from source: task vars 13731 1727203873.10243: variable 'controller_profile' from source: play vars 13731 1727203873.10317: variable 'controller_profile' from source: play vars 13731 1727203873.10325: variable 'controller_device' from source: play vars 13731 1727203873.10396: variable 'controller_device' from source: play vars 13731 1727203873.10404: variable 'dhcp_interface1' from source: play vars 13731 1727203873.10452: variable 'dhcp_interface1' from source: play vars 13731 1727203873.10460: variable 'port1_profile' from source: play vars 13731 1727203873.10532: variable 'port1_profile' from source: play vars 13731 1727203873.10540: variable 'dhcp_interface1' from source: play vars 13731 1727203873.10610: variable 'dhcp_interface1' from source: play vars 13731 1727203873.10619: variable 'controller_profile' from source: play vars 13731 1727203873.10687: variable 'controller_profile' from source: play vars 13731 1727203873.10694: variable 'port2_profile' from source: play vars 13731 1727203873.10765: variable 'port2_profile' from source: play vars 13731 1727203873.10772: variable 'dhcp_interface2' from source: play vars 13731 1727203873.10841: variable 'dhcp_interface2' from source: play vars 13731 1727203873.10849: variable 'controller_profile' from source: play vars 13731 1727203873.10917: variable 'controller_profile' from source: play vars 13731 1727203873.11072: variable '__network_packages_default_wireless' from source: role '' defaults 13731 1727203873.11076: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203873.11581: variable 'network_connections' from source: task vars 13731 1727203873.11584: variable 'controller_profile' from source: play vars 13731 1727203873.11587: variable 'controller_profile' from source: play vars 13731 1727203873.11589: variable 'controller_device' from source: play vars 13731 1727203873.11591: variable 'controller_device' from source: play vars 13731 1727203873.11593: variable 'dhcp_interface1' from source: play vars 13731 1727203873.11640: variable 'dhcp_interface1' from source: play vars 13731 1727203873.11647: variable 'port1_profile' from source: play vars 13731 1727203873.11709: variable 'port1_profile' from source: play vars 13731 1727203873.11715: variable 'dhcp_interface1' from source: play vars 13731 1727203873.11779: variable 'dhcp_interface1' from source: play vars 13731 1727203873.11785: variable 'controller_profile' from source: play vars 13731 1727203873.11844: variable 'controller_profile' from source: play vars 13731 1727203873.11856: variable 'port2_profile' from source: play vars 13731 1727203873.11932: variable 'port2_profile' from source: play vars 13731 1727203873.11938: variable 'dhcp_interface2' from source: play vars 13731 1727203873.12016: variable 'dhcp_interface2' from source: play vars 13731 1727203873.12023: variable 'controller_profile' from source: play vars 13731 1727203873.12101: variable 'controller_profile' from source: play vars 13731 1727203873.12121: variable '__network_packages_default_team' from source: role '' defaults 13731 1727203873.12205: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203873.12413: variable 'network_connections' from source: task vars 13731 1727203873.12416: variable 'controller_profile' from source: play vars 13731 1727203873.12464: variable 'controller_profile' from source: play vars 13731 1727203873.12467: variable 'controller_device' from source: play vars 13731 1727203873.12514: variable 'controller_device' from source: play vars 13731 1727203873.12520: variable 'dhcp_interface1' from source: play vars 13731 1727203873.12566: variable 'dhcp_interface1' from source: play vars 13731 1727203873.12573: variable 'port1_profile' from source: play vars 13731 1727203873.12620: variable 'port1_profile' from source: play vars 13731 1727203873.12626: variable 'dhcp_interface1' from source: play vars 13731 1727203873.12670: variable 'dhcp_interface1' from source: play vars 13731 1727203873.12675: variable 'controller_profile' from source: play vars 13731 1727203873.12723: variable 'controller_profile' from source: play vars 13731 1727203873.12729: variable 'port2_profile' from source: play vars 13731 1727203873.12773: variable 'port2_profile' from source: play vars 13731 1727203873.12780: variable 'dhcp_interface2' from source: play vars 13731 1727203873.12828: variable 'dhcp_interface2' from source: play vars 13731 1727203873.12834: variable 'controller_profile' from source: play vars 13731 1727203873.12879: variable 'controller_profile' from source: play vars 13731 1727203873.12929: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203873.12970: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203873.12980: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203873.13020: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203873.13154: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13731 1727203873.13467: variable 'network_connections' from source: task vars 13731 1727203873.13470: variable 'controller_profile' from source: play vars 13731 1727203873.13514: variable 'controller_profile' from source: play vars 13731 1727203873.13520: variable 'controller_device' from source: play vars 13731 1727203873.13567: variable 'controller_device' from source: play vars 13731 1727203873.13570: variable 'dhcp_interface1' from source: play vars 13731 1727203873.13611: variable 'dhcp_interface1' from source: play vars 13731 1727203873.13617: variable 'port1_profile' from source: play vars 13731 1727203873.13657: variable 'port1_profile' from source: play vars 13731 1727203873.13663: variable 'dhcp_interface1' from source: play vars 13731 1727203873.13707: variable 'dhcp_interface1' from source: play vars 13731 1727203873.13712: variable 'controller_profile' from source: play vars 13731 1727203873.13752: variable 'controller_profile' from source: play vars 13731 1727203873.13760: variable 'port2_profile' from source: play vars 13731 1727203873.13803: variable 'port2_profile' from source: play vars 13731 1727203873.13808: variable 'dhcp_interface2' from source: play vars 13731 1727203873.13848: variable 'dhcp_interface2' from source: play vars 13731 1727203873.13855: variable 'controller_profile' from source: play vars 13731 1727203873.13899: variable 'controller_profile' from source: play vars 13731 1727203873.13906: variable 'ansible_distribution' from source: facts 13731 1727203873.13909: variable '__network_rh_distros' from source: role '' defaults 13731 1727203873.13914: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.13934: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13731 1727203873.14043: variable 'ansible_distribution' from source: facts 13731 1727203873.14046: variable '__network_rh_distros' from source: role '' defaults 13731 1727203873.14051: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.14063: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13731 1727203873.14232: variable 'ansible_distribution' from source: facts 13731 1727203873.14235: variable '__network_rh_distros' from source: role '' defaults 13731 1727203873.14238: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.14240: variable 'network_provider' from source: set_fact 13731 1727203873.14285: variable 'ansible_facts' from source: unknown 13731 1727203873.14999: Evaluated conditional (not network_packages is subset(ansible_facts.packages.keys())): False 13731 1727203873.15002: when evaluation is False, skipping this task 13731 1727203873.15003: _execute() done 13731 1727203873.15005: dumping result to json 13731 1727203873.15007: done dumping result, returning 13731 1727203873.15009: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install packages [028d2410-947f-82dc-c122-000000000a36] 13731 1727203873.15010: sending task result for task 028d2410-947f-82dc-c122-000000000a36 13731 1727203873.15080: done sending task result for task 028d2410-947f-82dc-c122-000000000a36 13731 1727203873.15083: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "not network_packages is subset(ansible_facts.packages.keys())", "skip_reason": "Conditional result was False" } 13731 1727203873.15134: no more pending results, returning what we have 13731 1727203873.15138: results queue empty 13731 1727203873.15139: checking for any_errors_fatal 13731 1727203873.15146: done checking for any_errors_fatal 13731 1727203873.15147: checking for max_fail_percentage 13731 1727203873.15149: done checking for max_fail_percentage 13731 1727203873.15150: checking to see if all hosts have failed and the running result is not ok 13731 1727203873.15150: done checking to see if all hosts have failed 13731 1727203873.15151: getting the remaining hosts for this loop 13731 1727203873.15153: done getting the remaining hosts for this loop 13731 1727203873.15156: getting the next task for host managed-node3 13731 1727203873.15166: done getting next task for host managed-node3 13731 1727203873.15170: ^ task is: TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13731 1727203873.15174: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203873.15194: getting variables 13731 1727203873.15196: in VariableManager get_vars() 13731 1727203873.15236: Calling all_inventory to load vars for managed-node3 13731 1727203873.15239: Calling groups_inventory to load vars for managed-node3 13731 1727203873.15241: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203873.15249: Calling all_plugins_play to load vars for managed-node3 13731 1727203873.15252: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203873.15254: Calling groups_plugins_play to load vars for managed-node3 13731 1727203873.16803: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203873.18511: done with get_vars() 13731 1727203873.18542: done getting variables 13731 1727203873.18605: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:85 Tuesday 24 September 2024 14:51:13 -0400 (0:00:00.138) 0:00:51.417 ***** 13731 1727203873.18651: entering _queue_task() for managed-node3/package 13731 1727203873.19036: worker is 1 (out of 1 available) 13731 1727203873.19164: exiting _queue_task() for managed-node3/package 13731 1727203873.19176: done queuing things up, now waiting for results queue to drain 13731 1727203873.19178: waiting for pending results... 13731 1727203873.19498: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13731 1727203873.19594: in run() - task 028d2410-947f-82dc-c122-000000000a37 13731 1727203873.19682: variable 'ansible_search_path' from source: unknown 13731 1727203873.19685: variable 'ansible_search_path' from source: unknown 13731 1727203873.19688: calling self._execute() 13731 1727203873.19762: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203873.19774: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203873.19789: variable 'omit' from source: magic vars 13731 1727203873.20187: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.20204: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203873.20338: variable 'network_state' from source: role '' defaults 13731 1727203873.20363: Evaluated conditional (network_state != {}): False 13731 1727203873.20371: when evaluation is False, skipping this task 13731 1727203873.20380: _execute() done 13731 1727203873.20386: dumping result to json 13731 1727203873.20393: done dumping result, returning 13731 1727203873.20403: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable [028d2410-947f-82dc-c122-000000000a37] 13731 1727203873.20413: sending task result for task 028d2410-947f-82dc-c122-000000000a37 13731 1727203873.20542: done sending task result for task 028d2410-947f-82dc-c122-000000000a37 13731 1727203873.20546: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203873.20615: no more pending results, returning what we have 13731 1727203873.20620: results queue empty 13731 1727203873.20621: checking for any_errors_fatal 13731 1727203873.20628: done checking for any_errors_fatal 13731 1727203873.20629: checking for max_fail_percentage 13731 1727203873.20631: done checking for max_fail_percentage 13731 1727203873.20632: checking to see if all hosts have failed and the running result is not ok 13731 1727203873.20632: done checking to see if all hosts have failed 13731 1727203873.20633: getting the remaining hosts for this loop 13731 1727203873.20635: done getting the remaining hosts for this loop 13731 1727203873.20638: getting the next task for host managed-node3 13731 1727203873.20648: done getting next task for host managed-node3 13731 1727203873.20653: ^ task is: TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13731 1727203873.20662: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203873.20687: getting variables 13731 1727203873.20689: in VariableManager get_vars() 13731 1727203873.20735: Calling all_inventory to load vars for managed-node3 13731 1727203873.20738: Calling groups_inventory to load vars for managed-node3 13731 1727203873.20740: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203873.20753: Calling all_plugins_play to load vars for managed-node3 13731 1727203873.20756: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203873.20762: Calling groups_plugins_play to load vars for managed-node3 13731 1727203873.22771: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203873.24375: done with get_vars() 13731 1727203873.24405: done getting variables 13731 1727203873.24475: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:96 Tuesday 24 September 2024 14:51:13 -0400 (0:00:00.058) 0:00:51.476 ***** 13731 1727203873.24513: entering _queue_task() for managed-node3/package 13731 1727203873.24940: worker is 1 (out of 1 available) 13731 1727203873.24952: exiting _queue_task() for managed-node3/package 13731 1727203873.24965: done queuing things up, now waiting for results queue to drain 13731 1727203873.24966: waiting for pending results... 13731 1727203873.25405: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13731 1727203873.25409: in run() - task 028d2410-947f-82dc-c122-000000000a38 13731 1727203873.25421: variable 'ansible_search_path' from source: unknown 13731 1727203873.25428: variable 'ansible_search_path' from source: unknown 13731 1727203873.25471: calling self._execute() 13731 1727203873.25580: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203873.25593: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203873.25612: variable 'omit' from source: magic vars 13731 1727203873.26003: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.26021: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203873.26162: variable 'network_state' from source: role '' defaults 13731 1727203873.26181: Evaluated conditional (network_state != {}): False 13731 1727203873.26191: when evaluation is False, skipping this task 13731 1727203873.26199: _execute() done 13731 1727203873.26207: dumping result to json 13731 1727203873.26215: done dumping result, returning 13731 1727203873.26226: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable [028d2410-947f-82dc-c122-000000000a38] 13731 1727203873.26264: sending task result for task 028d2410-947f-82dc-c122-000000000a38 skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203873.26419: no more pending results, returning what we have 13731 1727203873.26424: results queue empty 13731 1727203873.26425: checking for any_errors_fatal 13731 1727203873.26434: done checking for any_errors_fatal 13731 1727203873.26435: checking for max_fail_percentage 13731 1727203873.26437: done checking for max_fail_percentage 13731 1727203873.26438: checking to see if all hosts have failed and the running result is not ok 13731 1727203873.26439: done checking to see if all hosts have failed 13731 1727203873.26439: getting the remaining hosts for this loop 13731 1727203873.26442: done getting the remaining hosts for this loop 13731 1727203873.26445: getting the next task for host managed-node3 13731 1727203873.26455: done getting next task for host managed-node3 13731 1727203873.26462: ^ task is: TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13731 1727203873.26468: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203873.26591: done sending task result for task 028d2410-947f-82dc-c122-000000000a38 13731 1727203873.26594: WORKER PROCESS EXITING 13731 1727203873.26610: getting variables 13731 1727203873.26612: in VariableManager get_vars() 13731 1727203873.26657: Calling all_inventory to load vars for managed-node3 13731 1727203873.26662: Calling groups_inventory to load vars for managed-node3 13731 1727203873.26665: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203873.26785: Calling all_plugins_play to load vars for managed-node3 13731 1727203873.26790: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203873.26794: Calling groups_plugins_play to load vars for managed-node3 13731 1727203873.28290: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203873.29922: done with get_vars() 13731 1727203873.29954: done getting variables 13731 1727203873.30022: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:109 Tuesday 24 September 2024 14:51:13 -0400 (0:00:00.055) 0:00:51.531 ***** 13731 1727203873.30072: entering _queue_task() for managed-node3/service 13731 1727203873.30513: worker is 1 (out of 1 available) 13731 1727203873.30526: exiting _queue_task() for managed-node3/service 13731 1727203873.30538: done queuing things up, now waiting for results queue to drain 13731 1727203873.30539: waiting for pending results... 13731 1727203873.30818: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13731 1727203873.31009: in run() - task 028d2410-947f-82dc-c122-000000000a39 13731 1727203873.31033: variable 'ansible_search_path' from source: unknown 13731 1727203873.31042: variable 'ansible_search_path' from source: unknown 13731 1727203873.31089: calling self._execute() 13731 1727203873.31230: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203873.31237: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203873.31240: variable 'omit' from source: magic vars 13731 1727203873.31636: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.31655: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203873.31882: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203873.32003: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203873.34008: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203873.34054: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203873.34085: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203873.34111: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203873.34131: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203873.34195: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.34220: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.34237: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.34265: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.34277: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.34313: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.34330: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.34346: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.34372: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.34384: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.34414: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.34432: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.34448: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.34473: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.34485: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.34780: variable 'network_connections' from source: task vars 13731 1727203873.34783: variable 'controller_profile' from source: play vars 13731 1727203873.34786: variable 'controller_profile' from source: play vars 13731 1727203873.34788: variable 'controller_device' from source: play vars 13731 1727203873.34823: variable 'controller_device' from source: play vars 13731 1727203873.34838: variable 'dhcp_interface1' from source: play vars 13731 1727203873.34914: variable 'dhcp_interface1' from source: play vars 13731 1727203873.34929: variable 'port1_profile' from source: play vars 13731 1727203873.35009: variable 'port1_profile' from source: play vars 13731 1727203873.35021: variable 'dhcp_interface1' from source: play vars 13731 1727203873.35081: variable 'dhcp_interface1' from source: play vars 13731 1727203873.35095: variable 'controller_profile' from source: play vars 13731 1727203873.35159: variable 'controller_profile' from source: play vars 13731 1727203873.35171: variable 'port2_profile' from source: play vars 13731 1727203873.35242: variable 'port2_profile' from source: play vars 13731 1727203873.35253: variable 'dhcp_interface2' from source: play vars 13731 1727203873.35317: variable 'dhcp_interface2' from source: play vars 13731 1727203873.35338: variable 'controller_profile' from source: play vars 13731 1727203873.35472: variable 'controller_profile' from source: play vars 13731 1727203873.35569: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203873.35769: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203873.35804: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203873.35827: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203873.35849: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203873.35892: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203873.35915: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203873.35932: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.35953: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203873.36009: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203873.36167: variable 'network_connections' from source: task vars 13731 1727203873.36170: variable 'controller_profile' from source: play vars 13731 1727203873.36220: variable 'controller_profile' from source: play vars 13731 1727203873.36226: variable 'controller_device' from source: play vars 13731 1727203873.36270: variable 'controller_device' from source: play vars 13731 1727203873.36278: variable 'dhcp_interface1' from source: play vars 13731 1727203873.36321: variable 'dhcp_interface1' from source: play vars 13731 1727203873.36328: variable 'port1_profile' from source: play vars 13731 1727203873.36370: variable 'port1_profile' from source: play vars 13731 1727203873.36377: variable 'dhcp_interface1' from source: play vars 13731 1727203873.36423: variable 'dhcp_interface1' from source: play vars 13731 1727203873.36426: variable 'controller_profile' from source: play vars 13731 1727203873.36466: variable 'controller_profile' from source: play vars 13731 1727203873.36472: variable 'port2_profile' from source: play vars 13731 1727203873.36514: variable 'port2_profile' from source: play vars 13731 1727203873.36522: variable 'dhcp_interface2' from source: play vars 13731 1727203873.36565: variable 'dhcp_interface2' from source: play vars 13731 1727203873.36570: variable 'controller_profile' from source: play vars 13731 1727203873.36612: variable 'controller_profile' from source: play vars 13731 1727203873.36642: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203873.36645: when evaluation is False, skipping this task 13731 1727203873.36648: _execute() done 13731 1727203873.36650: dumping result to json 13731 1727203873.36653: done dumping result, returning 13731 1727203873.36662: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000a39] 13731 1727203873.36668: sending task result for task 028d2410-947f-82dc-c122-000000000a39 13731 1727203873.36760: done sending task result for task 028d2410-947f-82dc-c122-000000000a39 13731 1727203873.36763: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203873.36816: no more pending results, returning what we have 13731 1727203873.36820: results queue empty 13731 1727203873.36821: checking for any_errors_fatal 13731 1727203873.36827: done checking for any_errors_fatal 13731 1727203873.36827: checking for max_fail_percentage 13731 1727203873.36829: done checking for max_fail_percentage 13731 1727203873.36830: checking to see if all hosts have failed and the running result is not ok 13731 1727203873.36831: done checking to see if all hosts have failed 13731 1727203873.36831: getting the remaining hosts for this loop 13731 1727203873.36833: done getting the remaining hosts for this loop 13731 1727203873.36836: getting the next task for host managed-node3 13731 1727203873.36844: done getting next task for host managed-node3 13731 1727203873.36847: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13731 1727203873.36852: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203873.36873: getting variables 13731 1727203873.36877: in VariableManager get_vars() 13731 1727203873.36917: Calling all_inventory to load vars for managed-node3 13731 1727203873.36920: Calling groups_inventory to load vars for managed-node3 13731 1727203873.36922: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203873.36931: Calling all_plugins_play to load vars for managed-node3 13731 1727203873.36934: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203873.36936: Calling groups_plugins_play to load vars for managed-node3 13731 1727203873.37897: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203873.39434: done with get_vars() 13731 1727203873.39465: done getting variables 13731 1727203873.39519: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start NetworkManager] ***** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 Tuesday 24 September 2024 14:51:13 -0400 (0:00:00.094) 0:00:51.626 ***** 13731 1727203873.39548: entering _queue_task() for managed-node3/service 13731 1727203873.39813: worker is 1 (out of 1 available) 13731 1727203873.39827: exiting _queue_task() for managed-node3/service 13731 1727203873.39840: done queuing things up, now waiting for results queue to drain 13731 1727203873.39842: waiting for pending results... 13731 1727203873.40027: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13731 1727203873.40129: in run() - task 028d2410-947f-82dc-c122-000000000a3a 13731 1727203873.40139: variable 'ansible_search_path' from source: unknown 13731 1727203873.40143: variable 'ansible_search_path' from source: unknown 13731 1727203873.40173: calling self._execute() 13731 1727203873.40249: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203873.40254: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203873.40263: variable 'omit' from source: magic vars 13731 1727203873.40541: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.40550: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203873.40666: variable 'network_provider' from source: set_fact 13731 1727203873.40669: variable 'network_state' from source: role '' defaults 13731 1727203873.40680: Evaluated conditional (network_provider == "nm" or network_state != {}): True 13731 1727203873.40686: variable 'omit' from source: magic vars 13731 1727203873.40737: variable 'omit' from source: magic vars 13731 1727203873.40760: variable 'network_service_name' from source: role '' defaults 13731 1727203873.40806: variable 'network_service_name' from source: role '' defaults 13731 1727203873.40878: variable '__network_provider_setup' from source: role '' defaults 13731 1727203873.40882: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203873.40926: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203873.40933: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203873.40980: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203873.41198: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203873.43226: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203873.43289: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203873.43324: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203873.43359: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203873.43384: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203873.43444: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.43478: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.43495: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.43521: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.43532: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.43569: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.43587: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.43603: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.43627: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.43638: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.43799: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13731 1727203873.43882: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.43900: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.43916: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.43939: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.43949: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.44017: variable 'ansible_python' from source: facts 13731 1727203873.44030: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13731 1727203873.44091: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203873.44144: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203873.44230: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.44247: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.44267: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.44295: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.44306: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.44338: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203873.44356: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203873.44376: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.44401: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203873.44413: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203873.44503: variable 'network_connections' from source: task vars 13731 1727203873.44509: variable 'controller_profile' from source: play vars 13731 1727203873.44562: variable 'controller_profile' from source: play vars 13731 1727203873.44574: variable 'controller_device' from source: play vars 13731 1727203873.44630: variable 'controller_device' from source: play vars 13731 1727203873.44639: variable 'dhcp_interface1' from source: play vars 13731 1727203873.44689: variable 'dhcp_interface1' from source: play vars 13731 1727203873.44699: variable 'port1_profile' from source: play vars 13731 1727203873.44753: variable 'port1_profile' from source: play vars 13731 1727203873.44764: variable 'dhcp_interface1' from source: play vars 13731 1727203873.44814: variable 'dhcp_interface1' from source: play vars 13731 1727203873.44822: variable 'controller_profile' from source: play vars 13731 1727203873.44878: variable 'controller_profile' from source: play vars 13731 1727203873.44887: variable 'port2_profile' from source: play vars 13731 1727203873.44935: variable 'port2_profile' from source: play vars 13731 1727203873.44944: variable 'dhcp_interface2' from source: play vars 13731 1727203873.45000: variable 'dhcp_interface2' from source: play vars 13731 1727203873.45009: variable 'controller_profile' from source: play vars 13731 1727203873.45059: variable 'controller_profile' from source: play vars 13731 1727203873.45133: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203873.45323: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203873.45381: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203873.45404: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203873.45444: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203873.45694: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203873.45697: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203873.45699: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203873.45702: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203873.45704: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203873.45920: variable 'network_connections' from source: task vars 13731 1727203873.45934: variable 'controller_profile' from source: play vars 13731 1727203873.46000: variable 'controller_profile' from source: play vars 13731 1727203873.46011: variable 'controller_device' from source: play vars 13731 1727203873.46085: variable 'controller_device' from source: play vars 13731 1727203873.46101: variable 'dhcp_interface1' from source: play vars 13731 1727203873.46157: variable 'dhcp_interface1' from source: play vars 13731 1727203873.46172: variable 'port1_profile' from source: play vars 13731 1727203873.46272: variable 'port1_profile' from source: play vars 13731 1727203873.46277: variable 'dhcp_interface1' from source: play vars 13731 1727203873.46322: variable 'dhcp_interface1' from source: play vars 13731 1727203873.46333: variable 'controller_profile' from source: play vars 13731 1727203873.46407: variable 'controller_profile' from source: play vars 13731 1727203873.46416: variable 'port2_profile' from source: play vars 13731 1727203873.46479: variable 'port2_profile' from source: play vars 13731 1727203873.46483: variable 'dhcp_interface2' from source: play vars 13731 1727203873.46545: variable 'dhcp_interface2' from source: play vars 13731 1727203873.46553: variable 'controller_profile' from source: play vars 13731 1727203873.46609: variable 'controller_profile' from source: play vars 13731 1727203873.46644: variable '__network_packages_default_wireless' from source: role '' defaults 13731 1727203873.46703: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203873.46893: variable 'network_connections' from source: task vars 13731 1727203873.46896: variable 'controller_profile' from source: play vars 13731 1727203873.46946: variable 'controller_profile' from source: play vars 13731 1727203873.46952: variable 'controller_device' from source: play vars 13731 1727203873.47003: variable 'controller_device' from source: play vars 13731 1727203873.47009: variable 'dhcp_interface1' from source: play vars 13731 1727203873.47071: variable 'dhcp_interface1' from source: play vars 13731 1727203873.47079: variable 'port1_profile' from source: play vars 13731 1727203873.47130: variable 'port1_profile' from source: play vars 13731 1727203873.47133: variable 'dhcp_interface1' from source: play vars 13731 1727203873.47182: variable 'dhcp_interface1' from source: play vars 13731 1727203873.47188: variable 'controller_profile' from source: play vars 13731 1727203873.47237: variable 'controller_profile' from source: play vars 13731 1727203873.47241: variable 'port2_profile' from source: play vars 13731 1727203873.47292: variable 'port2_profile' from source: play vars 13731 1727203873.47298: variable 'dhcp_interface2' from source: play vars 13731 1727203873.47346: variable 'dhcp_interface2' from source: play vars 13731 1727203873.47354: variable 'controller_profile' from source: play vars 13731 1727203873.47402: variable 'controller_profile' from source: play vars 13731 1727203873.47421: variable '__network_packages_default_team' from source: role '' defaults 13731 1727203873.47480: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203873.47659: variable 'network_connections' from source: task vars 13731 1727203873.47666: variable 'controller_profile' from source: play vars 13731 1727203873.47719: variable 'controller_profile' from source: play vars 13731 1727203873.47724: variable 'controller_device' from source: play vars 13731 1727203873.47773: variable 'controller_device' from source: play vars 13731 1727203873.47781: variable 'dhcp_interface1' from source: play vars 13731 1727203873.47838: variable 'dhcp_interface1' from source: play vars 13731 1727203873.47845: variable 'port1_profile' from source: play vars 13731 1727203873.47901: variable 'port1_profile' from source: play vars 13731 1727203873.47904: variable 'dhcp_interface1' from source: play vars 13731 1727203873.47950: variable 'dhcp_interface1' from source: play vars 13731 1727203873.47955: variable 'controller_profile' from source: play vars 13731 1727203873.48007: variable 'controller_profile' from source: play vars 13731 1727203873.48016: variable 'port2_profile' from source: play vars 13731 1727203873.48061: variable 'port2_profile' from source: play vars 13731 1727203873.48069: variable 'dhcp_interface2' from source: play vars 13731 1727203873.48119: variable 'dhcp_interface2' from source: play vars 13731 1727203873.48126: variable 'controller_profile' from source: play vars 13731 1727203873.48177: variable 'controller_profile' from source: play vars 13731 1727203873.48224: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203873.48266: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203873.48272: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203873.48315: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203873.48451: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13731 1727203873.48750: variable 'network_connections' from source: task vars 13731 1727203873.48755: variable 'controller_profile' from source: play vars 13731 1727203873.48802: variable 'controller_profile' from source: play vars 13731 1727203873.48808: variable 'controller_device' from source: play vars 13731 1727203873.48848: variable 'controller_device' from source: play vars 13731 1727203873.48854: variable 'dhcp_interface1' from source: play vars 13731 1727203873.48903: variable 'dhcp_interface1' from source: play vars 13731 1727203873.48909: variable 'port1_profile' from source: play vars 13731 1727203873.48949: variable 'port1_profile' from source: play vars 13731 1727203873.48955: variable 'dhcp_interface1' from source: play vars 13731 1727203873.49002: variable 'dhcp_interface1' from source: play vars 13731 1727203873.49008: variable 'controller_profile' from source: play vars 13731 1727203873.49048: variable 'controller_profile' from source: play vars 13731 1727203873.49055: variable 'port2_profile' from source: play vars 13731 1727203873.49102: variable 'port2_profile' from source: play vars 13731 1727203873.49105: variable 'dhcp_interface2' from source: play vars 13731 1727203873.49147: variable 'dhcp_interface2' from source: play vars 13731 1727203873.49152: variable 'controller_profile' from source: play vars 13731 1727203873.49200: variable 'controller_profile' from source: play vars 13731 1727203873.49204: variable 'ansible_distribution' from source: facts 13731 1727203873.49206: variable '__network_rh_distros' from source: role '' defaults 13731 1727203873.49212: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.49233: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13731 1727203873.49366: variable 'ansible_distribution' from source: facts 13731 1727203873.49369: variable '__network_rh_distros' from source: role '' defaults 13731 1727203873.49374: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.49387: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13731 1727203873.49501: variable 'ansible_distribution' from source: facts 13731 1727203873.49504: variable '__network_rh_distros' from source: role '' defaults 13731 1727203873.49508: variable 'ansible_distribution_major_version' from source: facts 13731 1727203873.49539: variable 'network_provider' from source: set_fact 13731 1727203873.49555: variable 'omit' from source: magic vars 13731 1727203873.49581: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203873.49602: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203873.49617: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203873.49635: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203873.49639: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203873.49666: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203873.49669: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203873.49672: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203873.49739: Set connection var ansible_pipelining to False 13731 1727203873.49742: Set connection var ansible_shell_type to sh 13731 1727203873.49751: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203873.49753: Set connection var ansible_connection to ssh 13731 1727203873.49757: Set connection var ansible_shell_executable to /bin/sh 13731 1727203873.49765: Set connection var ansible_timeout to 10 13731 1727203873.49784: variable 'ansible_shell_executable' from source: unknown 13731 1727203873.49787: variable 'ansible_connection' from source: unknown 13731 1727203873.49789: variable 'ansible_module_compression' from source: unknown 13731 1727203873.49791: variable 'ansible_shell_type' from source: unknown 13731 1727203873.49793: variable 'ansible_shell_executable' from source: unknown 13731 1727203873.49797: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203873.49801: variable 'ansible_pipelining' from source: unknown 13731 1727203873.49803: variable 'ansible_timeout' from source: unknown 13731 1727203873.49807: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203873.49885: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203873.49894: variable 'omit' from source: magic vars 13731 1727203873.49899: starting attempt loop 13731 1727203873.49902: running the handler 13731 1727203873.49954: variable 'ansible_facts' from source: unknown 13731 1727203873.50531: _low_level_execute_command(): starting 13731 1727203873.50536: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203873.51041: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203873.51045: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.51050: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203873.51053: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.51105: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203873.51109: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203873.51160: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203873.52843: stdout chunk (state=3): >>>/root <<< 13731 1727203873.52937: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203873.52967: stderr chunk (state=3): >>><<< 13731 1727203873.52978: stdout chunk (state=3): >>><<< 13731 1727203873.52997: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203873.53008: _low_level_execute_command(): starting 13731 1727203873.53014: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597 `" && echo ansible-tmp-1727203873.5299711-17759-73820572263597="` echo /root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597 `" ) && sleep 0' 13731 1727203873.53474: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203873.53479: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203873.53481: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.53483: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203873.53485: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203873.53488: stderr chunk (state=3): >>>debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.53538: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203873.53541: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203873.53545: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203873.53586: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203873.55451: stdout chunk (state=3): >>>ansible-tmp-1727203873.5299711-17759-73820572263597=/root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597 <<< 13731 1727203873.55554: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203873.55585: stderr chunk (state=3): >>><<< 13731 1727203873.55588: stdout chunk (state=3): >>><<< 13731 1727203873.55605: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203873.5299711-17759-73820572263597=/root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203873.55630: variable 'ansible_module_compression' from source: unknown 13731 1727203873.55673: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.systemd-ZIP_DEFLATED 13731 1727203873.55725: variable 'ansible_facts' from source: unknown 13731 1727203873.55860: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/AnsiballZ_systemd.py 13731 1727203873.55971: Sending initial data 13731 1727203873.55977: Sent initial data (155 bytes) 13731 1727203873.56432: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203873.56435: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203873.56438: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.56440: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203873.56442: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.56499: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203873.56502: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203873.56506: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203873.56539: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203873.58097: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" <<< 13731 1727203873.58104: stderr chunk (state=3): >>>debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203873.58131: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203873.58166: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpmc53xxwq /root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/AnsiballZ_systemd.py <<< 13731 1727203873.58169: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/AnsiballZ_systemd.py" <<< 13731 1727203873.58206: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpmc53xxwq" to remote "/root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/AnsiballZ_systemd.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/AnsiballZ_systemd.py" <<< 13731 1727203873.59228: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203873.59269: stderr chunk (state=3): >>><<< 13731 1727203873.59273: stdout chunk (state=3): >>><<< 13731 1727203873.59316: done transferring module to remote 13731 1727203873.59325: _low_level_execute_command(): starting 13731 1727203873.59329: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/ /root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/AnsiballZ_systemd.py && sleep 0' 13731 1727203873.59752: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203873.59760: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203873.59780: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.59784: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203873.59794: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.59856: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203873.59858: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203873.59861: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203873.59886: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203873.61633: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203873.61659: stderr chunk (state=3): >>><<< 13731 1727203873.61662: stdout chunk (state=3): >>><<< 13731 1727203873.61683: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203873.61687: _low_level_execute_command(): starting 13731 1727203873.61691: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/AnsiballZ_systemd.py && sleep 0' 13731 1727203873.62145: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203873.62148: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.62151: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203873.62153: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203873.62155: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203873.62201: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203873.62205: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203873.62207: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203873.62253: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203873.91312: stdout chunk (state=3): >>> {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "RestartMode": "normal", "NotifyAccess": "none", "RestartUSec": "100ms", "RestartSteps": "0", "RestartMaxDelayUSec": "infinity", "RestartUSecNext": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "704", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "FileDescriptorStorePreserve": "restart", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainStartTimestampMonotonic": "28990148", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainHandoffTimestampMonotonic": "29005881", "ExecMainPID": "704", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "2938", "MemoryCurrent": "10350592", "MemoryPeak": "13586432", "MemorySwapCurrent": "0", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryAvailable": "3308949504", "EffectiveMemoryMax": "3702865920", "EffectiveMemoryHigh": "3702865920", "CPUUsageNSec": "841317000", "TasksCurrent": "4", "EffectiveTasksMax": "22362", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultStartupMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "StartupMemoryLow": "0", "MemoryHigh": "infinity", "StartupMemoryHigh": "infinity", "MemoryMax": "infinity", "StartupMemoryMax": "infinity", "MemorySwapMax": "infinity", "StartupMemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22362", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "MemoryPressureWatch": "auto", "MemoryPressureThresholdUSec": "200ms", "CoredumpRe<<< 13731 1727203873.91325: stdout chunk (state=3): >>>ceive": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "RootEphemeral": "no", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "SetLoginEnvironment": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "MemoryKSM": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "system.slice dbus.socket sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "cloud-init.service shutdown.target NetworkManager-wait-online.service multi-user.target network.target", "After": "dbus-broker.service systemd-journald.socket network-pre.target basic.target cloud-init-local.service dbus.socket system.slice sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:49:45 EDT", "StateChangeTimestampMonotonic": "362725592", "InactiveExitTimestamp": "Tue 2024-09-24 14:44:11 EDT", "InactiveExitTimestampMonotonic": "28990654", "ActiveEnterTimestamp": "Tue 2024-09-24 14:44:12 EDT", "ActiveEnterTimestampMonotonic": "29769382", "ActiveExitTimestampMonotonic": "0", "InactiveEnterTimestampMonotonic": "0", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "SurviveFinalKillSignal": "no", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ConditionTimestampMonotonic": "28989295", "AssertTimestamp": "Tue 2024-09-24 14:44:11 EDT", "AssertTimestampMonotonic": "28989297", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "70a845f8a1964db89963090ed497f47f", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 13731 1727203873.93290: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203873.93294: stdout chunk (state=3): >>><<< 13731 1727203873.93297: stderr chunk (state=3): >>><<< 13731 1727203873.93301: _low_level_execute_command() done: rc=0, stdout= {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "RestartMode": "normal", "NotifyAccess": "none", "RestartUSec": "100ms", "RestartSteps": "0", "RestartMaxDelayUSec": "infinity", "RestartUSecNext": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "704", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "FileDescriptorStorePreserve": "restart", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainStartTimestampMonotonic": "28990148", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainHandoffTimestampMonotonic": "29005881", "ExecMainPID": "704", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "2938", "MemoryCurrent": "10350592", "MemoryPeak": "13586432", "MemorySwapCurrent": "0", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryAvailable": "3308949504", "EffectiveMemoryMax": "3702865920", "EffectiveMemoryHigh": "3702865920", "CPUUsageNSec": "841317000", "TasksCurrent": "4", "EffectiveTasksMax": "22362", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultStartupMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "StartupMemoryLow": "0", "MemoryHigh": "infinity", "StartupMemoryHigh": "infinity", "MemoryMax": "infinity", "StartupMemoryMax": "infinity", "MemorySwapMax": "infinity", "StartupMemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22362", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "MemoryPressureWatch": "auto", "MemoryPressureThresholdUSec": "200ms", "CoredumpReceive": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "RootEphemeral": "no", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "SetLoginEnvironment": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "MemoryKSM": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "system.slice dbus.socket sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "cloud-init.service shutdown.target NetworkManager-wait-online.service multi-user.target network.target", "After": "dbus-broker.service systemd-journald.socket network-pre.target basic.target cloud-init-local.service dbus.socket system.slice sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:49:45 EDT", "StateChangeTimestampMonotonic": "362725592", "InactiveExitTimestamp": "Tue 2024-09-24 14:44:11 EDT", "InactiveExitTimestampMonotonic": "28990654", "ActiveEnterTimestamp": "Tue 2024-09-24 14:44:12 EDT", "ActiveEnterTimestampMonotonic": "29769382", "ActiveExitTimestampMonotonic": "0", "InactiveEnterTimestampMonotonic": "0", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "SurviveFinalKillSignal": "no", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ConditionTimestampMonotonic": "28989295", "AssertTimestamp": "Tue 2024-09-24 14:44:11 EDT", "AssertTimestampMonotonic": "28989297", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "70a845f8a1964db89963090ed497f47f", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203873.93504: done with _execute_module (ansible.legacy.systemd, {'name': 'NetworkManager', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203873.93761: _low_level_execute_command(): starting 13731 1727203873.93765: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203873.5299711-17759-73820572263597/ > /dev/null 2>&1 && sleep 0' 13731 1727203873.94365: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203873.94384: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203873.94416: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203873.94519: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203873.94540: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203873.94608: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203873.96512: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203873.96822: stderr chunk (state=3): >>><<< 13731 1727203873.96825: stdout chunk (state=3): >>><<< 13731 1727203873.96828: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203873.96830: handler run complete 13731 1727203873.96832: attempt loop complete, returning result 13731 1727203873.96834: _execute() done 13731 1727203873.96836: dumping result to json 13731 1727203873.97149: done dumping result, returning 13731 1727203873.97152: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager [028d2410-947f-82dc-c122-000000000a3a] 13731 1727203873.97154: sending task result for task 028d2410-947f-82dc-c122-000000000a3a ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203873.97742: no more pending results, returning what we have 13731 1727203873.97747: results queue empty 13731 1727203873.97748: checking for any_errors_fatal 13731 1727203873.97756: done checking for any_errors_fatal 13731 1727203873.97757: checking for max_fail_percentage 13731 1727203873.97762: done checking for max_fail_percentage 13731 1727203873.97763: checking to see if all hosts have failed and the running result is not ok 13731 1727203873.97764: done checking to see if all hosts have failed 13731 1727203873.97764: getting the remaining hosts for this loop 13731 1727203873.97767: done getting the remaining hosts for this loop 13731 1727203873.97770: getting the next task for host managed-node3 13731 1727203873.97780: done getting next task for host managed-node3 13731 1727203873.97784: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13731 1727203873.97985: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203873.98001: getting variables 13731 1727203873.98003: in VariableManager get_vars() 13731 1727203873.98287: Calling all_inventory to load vars for managed-node3 13731 1727203873.98291: Calling groups_inventory to load vars for managed-node3 13731 1727203873.98295: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203873.98306: Calling all_plugins_play to load vars for managed-node3 13731 1727203873.98309: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203873.98312: Calling groups_plugins_play to load vars for managed-node3 13731 1727203873.99030: done sending task result for task 028d2410-947f-82dc-c122-000000000a3a 13731 1727203873.99034: WORKER PROCESS EXITING 13731 1727203874.00803: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203874.02661: done with get_vars() 13731 1727203874.02690: done getting variables 13731 1727203874.02756: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start wpa_supplicant] ***** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:133 Tuesday 24 September 2024 14:51:14 -0400 (0:00:00.632) 0:00:52.258 ***** 13731 1727203874.02801: entering _queue_task() for managed-node3/service 13731 1727203874.03183: worker is 1 (out of 1 available) 13731 1727203874.03195: exiting _queue_task() for managed-node3/service 13731 1727203874.03207: done queuing things up, now waiting for results queue to drain 13731 1727203874.03209: waiting for pending results... 13731 1727203874.03423: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13731 1727203874.03536: in run() - task 028d2410-947f-82dc-c122-000000000a3b 13731 1727203874.03547: variable 'ansible_search_path' from source: unknown 13731 1727203874.03552: variable 'ansible_search_path' from source: unknown 13731 1727203874.03587: calling self._execute() 13731 1727203874.03665: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203874.03670: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203874.03681: variable 'omit' from source: magic vars 13731 1727203874.03959: variable 'ansible_distribution_major_version' from source: facts 13731 1727203874.03971: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203874.04052: variable 'network_provider' from source: set_fact 13731 1727203874.04055: Evaluated conditional (network_provider == "nm"): True 13731 1727203874.04122: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203874.04186: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203874.04305: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203874.05881: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203874.05928: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203874.05964: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203874.06008: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203874.06051: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203874.06267: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203874.06271: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203874.06274: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203874.06297: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203874.06319: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203874.06369: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203874.06407: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203874.06480: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203874.06484: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203874.06504: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203874.06549: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203874.06578: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203874.06614: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203874.06658: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203874.06680: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203874.06844: variable 'network_connections' from source: task vars 13731 1727203874.06901: variable 'controller_profile' from source: play vars 13731 1727203874.06925: variable 'controller_profile' from source: play vars 13731 1727203874.06942: variable 'controller_device' from source: play vars 13731 1727203874.07004: variable 'controller_device' from source: play vars 13731 1727203874.07012: variable 'dhcp_interface1' from source: play vars 13731 1727203874.07062: variable 'dhcp_interface1' from source: play vars 13731 1727203874.07066: variable 'port1_profile' from source: play vars 13731 1727203874.07111: variable 'port1_profile' from source: play vars 13731 1727203874.07117: variable 'dhcp_interface1' from source: play vars 13731 1727203874.07157: variable 'dhcp_interface1' from source: play vars 13731 1727203874.07163: variable 'controller_profile' from source: play vars 13731 1727203874.07216: variable 'controller_profile' from source: play vars 13731 1727203874.07221: variable 'port2_profile' from source: play vars 13731 1727203874.07268: variable 'port2_profile' from source: play vars 13731 1727203874.07271: variable 'dhcp_interface2' from source: play vars 13731 1727203874.07313: variable 'dhcp_interface2' from source: play vars 13731 1727203874.07320: variable 'controller_profile' from source: play vars 13731 1727203874.07362: variable 'controller_profile' from source: play vars 13731 1727203874.07412: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203874.07527: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203874.07555: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203874.07579: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203874.07603: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203874.07634: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203874.07649: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203874.07671: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203874.07690: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203874.07731: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203874.07890: variable 'network_connections' from source: task vars 13731 1727203874.07893: variable 'controller_profile' from source: play vars 13731 1727203874.07937: variable 'controller_profile' from source: play vars 13731 1727203874.07943: variable 'controller_device' from source: play vars 13731 1727203874.07985: variable 'controller_device' from source: play vars 13731 1727203874.07991: variable 'dhcp_interface1' from source: play vars 13731 1727203874.08033: variable 'dhcp_interface1' from source: play vars 13731 1727203874.08039: variable 'port1_profile' from source: play vars 13731 1727203874.08082: variable 'port1_profile' from source: play vars 13731 1727203874.08088: variable 'dhcp_interface1' from source: play vars 13731 1727203874.08129: variable 'dhcp_interface1' from source: play vars 13731 1727203874.08133: variable 'controller_profile' from source: play vars 13731 1727203874.08177: variable 'controller_profile' from source: play vars 13731 1727203874.08183: variable 'port2_profile' from source: play vars 13731 1727203874.08222: variable 'port2_profile' from source: play vars 13731 1727203874.08228: variable 'dhcp_interface2' from source: play vars 13731 1727203874.08272: variable 'dhcp_interface2' from source: play vars 13731 1727203874.08278: variable 'controller_profile' from source: play vars 13731 1727203874.08318: variable 'controller_profile' from source: play vars 13731 1727203874.08348: Evaluated conditional (__network_wpa_supplicant_required): False 13731 1727203874.08353: when evaluation is False, skipping this task 13731 1727203874.08356: _execute() done 13731 1727203874.08360: dumping result to json 13731 1727203874.08363: done dumping result, returning 13731 1727203874.08365: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant [028d2410-947f-82dc-c122-000000000a3b] 13731 1727203874.08375: sending task result for task 028d2410-947f-82dc-c122-000000000a3b 13731 1727203874.08460: done sending task result for task 028d2410-947f-82dc-c122-000000000a3b 13731 1727203874.08463: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wpa_supplicant_required", "skip_reason": "Conditional result was False" } 13731 1727203874.08527: no more pending results, returning what we have 13731 1727203874.08531: results queue empty 13731 1727203874.08532: checking for any_errors_fatal 13731 1727203874.08552: done checking for any_errors_fatal 13731 1727203874.08552: checking for max_fail_percentage 13731 1727203874.08555: done checking for max_fail_percentage 13731 1727203874.08555: checking to see if all hosts have failed and the running result is not ok 13731 1727203874.08556: done checking to see if all hosts have failed 13731 1727203874.08557: getting the remaining hosts for this loop 13731 1727203874.08561: done getting the remaining hosts for this loop 13731 1727203874.08564: getting the next task for host managed-node3 13731 1727203874.08572: done getting next task for host managed-node3 13731 1727203874.08579: ^ task is: TASK: fedora.linux_system_roles.network : Enable network service 13731 1727203874.08589: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203874.08609: getting variables 13731 1727203874.08610: in VariableManager get_vars() 13731 1727203874.08647: Calling all_inventory to load vars for managed-node3 13731 1727203874.08650: Calling groups_inventory to load vars for managed-node3 13731 1727203874.08653: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203874.08664: Calling all_plugins_play to load vars for managed-node3 13731 1727203874.08667: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203874.08669: Calling groups_plugins_play to load vars for managed-node3 13731 1727203874.09815: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203874.11184: done with get_vars() 13731 1727203874.11206: done getting variables 13731 1727203874.11249: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable network service] ************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:142 Tuesday 24 September 2024 14:51:14 -0400 (0:00:00.084) 0:00:52.343 ***** 13731 1727203874.11278: entering _queue_task() for managed-node3/service 13731 1727203874.11534: worker is 1 (out of 1 available) 13731 1727203874.11549: exiting _queue_task() for managed-node3/service 13731 1727203874.11564: done queuing things up, now waiting for results queue to drain 13731 1727203874.11566: waiting for pending results... 13731 1727203874.11753: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable network service 13731 1727203874.11862: in run() - task 028d2410-947f-82dc-c122-000000000a3c 13731 1727203874.11878: variable 'ansible_search_path' from source: unknown 13731 1727203874.11882: variable 'ansible_search_path' from source: unknown 13731 1727203874.11912: calling self._execute() 13731 1727203874.11990: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203874.11994: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203874.12004: variable 'omit' from source: magic vars 13731 1727203874.12285: variable 'ansible_distribution_major_version' from source: facts 13731 1727203874.12295: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203874.12378: variable 'network_provider' from source: set_fact 13731 1727203874.12382: Evaluated conditional (network_provider == "initscripts"): False 13731 1727203874.12385: when evaluation is False, skipping this task 13731 1727203874.12387: _execute() done 13731 1727203874.12390: dumping result to json 13731 1727203874.12394: done dumping result, returning 13731 1727203874.12400: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable network service [028d2410-947f-82dc-c122-000000000a3c] 13731 1727203874.12406: sending task result for task 028d2410-947f-82dc-c122-000000000a3c 13731 1727203874.12493: done sending task result for task 028d2410-947f-82dc-c122-000000000a3c 13731 1727203874.12496: WORKER PROCESS EXITING skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203874.12540: no more pending results, returning what we have 13731 1727203874.12544: results queue empty 13731 1727203874.12545: checking for any_errors_fatal 13731 1727203874.12557: done checking for any_errors_fatal 13731 1727203874.12557: checking for max_fail_percentage 13731 1727203874.12559: done checking for max_fail_percentage 13731 1727203874.12560: checking to see if all hosts have failed and the running result is not ok 13731 1727203874.12561: done checking to see if all hosts have failed 13731 1727203874.12561: getting the remaining hosts for this loop 13731 1727203874.12563: done getting the remaining hosts for this loop 13731 1727203874.12566: getting the next task for host managed-node3 13731 1727203874.12575: done getting next task for host managed-node3 13731 1727203874.12580: ^ task is: TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13731 1727203874.12585: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=19, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203874.12608: getting variables 13731 1727203874.12609: in VariableManager get_vars() 13731 1727203874.12648: Calling all_inventory to load vars for managed-node3 13731 1727203874.12652: Calling groups_inventory to load vars for managed-node3 13731 1727203874.12654: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203874.12663: Calling all_plugins_play to load vars for managed-node3 13731 1727203874.12666: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203874.12668: Calling groups_plugins_play to load vars for managed-node3 13731 1727203874.13974: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203874.14857: done with get_vars() 13731 1727203874.14878: done getting variables 13731 1727203874.14920: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ensure initscripts network file dependency is present] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:150 Tuesday 24 September 2024 14:51:14 -0400 (0:00:00.036) 0:00:52.380 ***** 13731 1727203874.14947: entering _queue_task() for managed-node3/copy 13731 1727203874.15198: worker is 1 (out of 1 available) 13731 1727203874.15212: exiting _queue_task() for managed-node3/copy 13731 1727203874.15224: done queuing things up, now waiting for results queue to drain 13731 1727203874.15226: waiting for pending results... 13731 1727203874.15409: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13731 1727203874.15520: in run() - task 028d2410-947f-82dc-c122-000000000a3d 13731 1727203874.15533: variable 'ansible_search_path' from source: unknown 13731 1727203874.15537: variable 'ansible_search_path' from source: unknown 13731 1727203874.15569: calling self._execute() 13731 1727203874.15646: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203874.15650: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203874.15657: variable 'omit' from source: magic vars 13731 1727203874.15932: variable 'ansible_distribution_major_version' from source: facts 13731 1727203874.15942: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203874.16027: variable 'network_provider' from source: set_fact 13731 1727203874.16031: Evaluated conditional (network_provider == "initscripts"): False 13731 1727203874.16034: when evaluation is False, skipping this task 13731 1727203874.16037: _execute() done 13731 1727203874.16039: dumping result to json 13731 1727203874.16041: done dumping result, returning 13731 1727203874.16050: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present [028d2410-947f-82dc-c122-000000000a3d] 13731 1727203874.16053: sending task result for task 028d2410-947f-82dc-c122-000000000a3d 13731 1727203874.16150: done sending task result for task 028d2410-947f-82dc-c122-000000000a3d 13731 1727203874.16154: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_provider == \"initscripts\"", "skip_reason": "Conditional result was False" } 13731 1727203874.16209: no more pending results, returning what we have 13731 1727203874.16214: results queue empty 13731 1727203874.16214: checking for any_errors_fatal 13731 1727203874.16222: done checking for any_errors_fatal 13731 1727203874.16222: checking for max_fail_percentage 13731 1727203874.16224: done checking for max_fail_percentage 13731 1727203874.16225: checking to see if all hosts have failed and the running result is not ok 13731 1727203874.16226: done checking to see if all hosts have failed 13731 1727203874.16227: getting the remaining hosts for this loop 13731 1727203874.16229: done getting the remaining hosts for this loop 13731 1727203874.16232: getting the next task for host managed-node3 13731 1727203874.16240: done getting next task for host managed-node3 13731 1727203874.16244: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13731 1727203874.16248: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=20, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203874.16277: getting variables 13731 1727203874.16279: in VariableManager get_vars() 13731 1727203874.16315: Calling all_inventory to load vars for managed-node3 13731 1727203874.16317: Calling groups_inventory to load vars for managed-node3 13731 1727203874.16319: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203874.16328: Calling all_plugins_play to load vars for managed-node3 13731 1727203874.16330: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203874.16333: Calling groups_plugins_play to load vars for managed-node3 13731 1727203874.17109: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203874.17969: done with get_vars() 13731 1727203874.17990: done getting variables TASK [fedora.linux_system_roles.network : Configure networking connection profiles] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 Tuesday 24 September 2024 14:51:14 -0400 (0:00:00.031) 0:00:52.411 ***** 13731 1727203874.18056: entering _queue_task() for managed-node3/fedora.linux_system_roles.network_connections 13731 1727203874.18315: worker is 1 (out of 1 available) 13731 1727203874.18329: exiting _queue_task() for managed-node3/fedora.linux_system_roles.network_connections 13731 1727203874.18342: done queuing things up, now waiting for results queue to drain 13731 1727203874.18344: waiting for pending results... 13731 1727203874.18545: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13731 1727203874.18666: in run() - task 028d2410-947f-82dc-c122-000000000a3e 13731 1727203874.18682: variable 'ansible_search_path' from source: unknown 13731 1727203874.18686: variable 'ansible_search_path' from source: unknown 13731 1727203874.18710: calling self._execute() 13731 1727203874.18791: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203874.18797: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203874.18806: variable 'omit' from source: magic vars 13731 1727203874.19077: variable 'ansible_distribution_major_version' from source: facts 13731 1727203874.19086: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203874.19091: variable 'omit' from source: magic vars 13731 1727203874.19145: variable 'omit' from source: magic vars 13731 1727203874.19264: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203874.21063: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203874.21115: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203874.21143: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203874.21170: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203874.21193: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203874.21381: variable 'network_provider' from source: set_fact 13731 1727203874.21420: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203874.21430: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203874.21455: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203874.21498: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203874.21510: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203874.21583: variable 'omit' from source: magic vars 13731 1727203874.21744: variable 'omit' from source: magic vars 13731 1727203874.21783: variable 'network_connections' from source: task vars 13731 1727203874.21794: variable 'controller_profile' from source: play vars 13731 1727203874.21865: variable 'controller_profile' from source: play vars 13731 1727203874.21869: variable 'controller_device' from source: play vars 13731 1727203874.21974: variable 'controller_device' from source: play vars 13731 1727203874.21980: variable 'dhcp_interface1' from source: play vars 13731 1727203874.21983: variable 'dhcp_interface1' from source: play vars 13731 1727203874.21992: variable 'port1_profile' from source: play vars 13731 1727203874.22047: variable 'port1_profile' from source: play vars 13731 1727203874.22054: variable 'dhcp_interface1' from source: play vars 13731 1727203874.22113: variable 'dhcp_interface1' from source: play vars 13731 1727203874.22119: variable 'controller_profile' from source: play vars 13731 1727203874.22179: variable 'controller_profile' from source: play vars 13731 1727203874.22190: variable 'port2_profile' from source: play vars 13731 1727203874.22238: variable 'port2_profile' from source: play vars 13731 1727203874.22256: variable 'dhcp_interface2' from source: play vars 13731 1727203874.22305: variable 'dhcp_interface2' from source: play vars 13731 1727203874.22311: variable 'controller_profile' from source: play vars 13731 1727203874.22425: variable 'controller_profile' from source: play vars 13731 1727203874.22552: variable 'omit' from source: magic vars 13731 1727203874.22560: variable '__lsr_ansible_managed' from source: task vars 13731 1727203874.22621: variable '__lsr_ansible_managed' from source: task vars 13731 1727203874.22806: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 13731 1727203874.23086: Loaded config def from plugin (lookup/template) 13731 1727203874.23089: Loading LookupModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/template.py 13731 1727203874.23091: File lookup term: get_ansible_managed.j2 13731 1727203874.23094: variable 'ansible_search_path' from source: unknown 13731 1727203874.23097: evaluation_path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks 13731 1727203874.23100: search_path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 13731 1727203874.23104: variable 'ansible_search_path' from source: unknown 13731 1727203874.33465: variable 'ansible_managed' from source: unknown 13731 1727203874.33556: variable 'omit' from source: magic vars 13731 1727203874.33580: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203874.33595: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203874.33606: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203874.33618: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203874.33625: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203874.33640: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203874.33647: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203874.33650: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203874.33715: Set connection var ansible_pipelining to False 13731 1727203874.33718: Set connection var ansible_shell_type to sh 13731 1727203874.33726: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203874.33737: Set connection var ansible_connection to ssh 13731 1727203874.33739: Set connection var ansible_shell_executable to /bin/sh 13731 1727203874.33741: Set connection var ansible_timeout to 10 13731 1727203874.33762: variable 'ansible_shell_executable' from source: unknown 13731 1727203874.33765: variable 'ansible_connection' from source: unknown 13731 1727203874.33769: variable 'ansible_module_compression' from source: unknown 13731 1727203874.33771: variable 'ansible_shell_type' from source: unknown 13731 1727203874.33774: variable 'ansible_shell_executable' from source: unknown 13731 1727203874.33777: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203874.33780: variable 'ansible_pipelining' from source: unknown 13731 1727203874.33782: variable 'ansible_timeout' from source: unknown 13731 1727203874.33784: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203874.33868: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203874.33878: variable 'omit' from source: magic vars 13731 1727203874.33884: starting attempt loop 13731 1727203874.33887: running the handler 13731 1727203874.33898: _low_level_execute_command(): starting 13731 1727203874.33903: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203874.34490: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203874.34610: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203874.34613: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203874.36275: stdout chunk (state=3): >>>/root <<< 13731 1727203874.36427: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203874.36431: stdout chunk (state=3): >>><<< 13731 1727203874.36433: stderr chunk (state=3): >>><<< 13731 1727203874.36451: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203874.36471: _low_level_execute_command(): starting 13731 1727203874.36485: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518 `" && echo ansible-tmp-1727203874.3645885-17791-192637604659518="` echo /root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518 `" ) && sleep 0' 13731 1727203874.37111: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203874.37206: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203874.37222: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203874.37243: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203874.37259: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203874.37270: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203874.37284: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203874.37349: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203874.37388: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203874.37413: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203874.37583: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203874.39426: stdout chunk (state=3): >>>ansible-tmp-1727203874.3645885-17791-192637604659518=/root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518 <<< 13731 1727203874.39626: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203874.39630: stdout chunk (state=3): >>><<< 13731 1727203874.39632: stderr chunk (state=3): >>><<< 13731 1727203874.39635: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203874.3645885-17791-192637604659518=/root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203874.39671: variable 'ansible_module_compression' from source: unknown 13731 1727203874.39781: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.network_connections-ZIP_DEFLATED 13731 1727203874.39896: variable 'ansible_facts' from source: unknown 13731 1727203874.40382: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/AnsiballZ_network_connections.py 13731 1727203874.40619: Sending initial data 13731 1727203874.40622: Sent initial data (168 bytes) 13731 1727203874.41923: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203874.42027: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203874.42045: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203874.42135: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203874.43610: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203874.43773: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203874.43818: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/AnsiballZ_network_connections.py" <<< 13731 1727203874.44080: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpbvrla7v8 /root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/AnsiballZ_network_connections.py <<< 13731 1727203874.44085: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpbvrla7v8" to remote "/root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/AnsiballZ_network_connections.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/AnsiballZ_network_connections.py" <<< 13731 1727203874.45532: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203874.45583: stderr chunk (state=3): >>><<< 13731 1727203874.45635: stdout chunk (state=3): >>><<< 13731 1727203874.45661: done transferring module to remote 13731 1727203874.45784: _low_level_execute_command(): starting 13731 1727203874.45788: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/ /root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/AnsiballZ_network_connections.py && sleep 0' 13731 1727203874.46943: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203874.47101: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203874.47318: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203874.47331: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203874.49094: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203874.49097: stdout chunk (state=3): >>><<< 13731 1727203874.49100: stderr chunk (state=3): >>><<< 13731 1727203874.49121: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203874.49132: _low_level_execute_command(): starting 13731 1727203874.49143: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/AnsiballZ_network_connections.py && sleep 0' 13731 1727203874.49783: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203874.49798: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203874.49811: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203874.49828: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203874.49882: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203874.49946: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203874.49968: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203874.50112: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203874.90860: stdout chunk (state=3): >>> {"changed": true, "warnings": [], "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 (not-active)\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "active-backup", "arp_interval": 60, "arp_ip_target": "192.0.2.128", "arp_validate": "none", "primary": "test1"}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "active-backup", "arp_interval": 60, "arp_ip_target": "192.0.2.128", "arp_validate": "none", "primary": "test1"}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed<<< 13731 1727203874.90937: stdout chunk (state=3): >>>\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} <<< 13731 1727203874.92903: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203874.92933: stderr chunk (state=3): >>><<< 13731 1727203874.92981: stdout chunk (state=3): >>><<< 13731 1727203874.92985: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "warnings": [], "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 (not-active)\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "active-backup", "arp_interval": 60, "arp_ip_target": "192.0.2.128", "arp_validate": "none", "primary": "test1"}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0", "state": "up", "type": "bond", "interface_name": "nm-bond", "bond": {"mode": "active-backup", "arp_interval": 60, "arp_ip_target": "192.0.2.128", "arp_validate": "none", "primary": "test1"}, "ip": {"route_metric4": 65535}}, {"name": "bond0.0", "state": "up", "type": "ethernet", "interface_name": "test1", "controller": "bond0"}, {"name": "bond0.1", "state": "up", "type": "ethernet", "interface_name": "test2", "controller": "bond0"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203874.93043: done with _execute_module (fedora.linux_system_roles.network_connections, {'provider': 'nm', 'connections': [{'name': 'bond0', 'state': 'up', 'type': 'bond', 'interface_name': 'nm-bond', 'bond': {'mode': 'active-backup', 'arp_interval': 60, 'arp_ip_target': '192.0.2.128', 'arp_validate': 'none', 'primary': 'test1'}, 'ip': {'route_metric4': 65535}}, {'name': 'bond0.0', 'state': 'up', 'type': 'ethernet', 'interface_name': 'test1', 'controller': 'bond0'}, {'name': 'bond0.1', 'state': 'up', 'type': 'ethernet', 'interface_name': 'test2', 'controller': 'bond0'}], '__header': '#\n# Ansible managed\n#\n# system_role:network\n', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'fedora.linux_system_roles.network_connections', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203874.93102: _low_level_execute_command(): starting 13731 1727203874.93105: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203874.3645885-17791-192637604659518/ > /dev/null 2>&1 && sleep 0' 13731 1727203874.93717: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203874.93827: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203874.93897: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203874.93908: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203874.95769: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203874.95783: stdout chunk (state=3): >>><<< 13731 1727203874.95796: stderr chunk (state=3): >>><<< 13731 1727203874.95818: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203874.95981: handler run complete 13731 1727203874.95984: attempt loop complete, returning result 13731 1727203874.95986: _execute() done 13731 1727203874.95988: dumping result to json 13731 1727203874.95990: done dumping result, returning 13731 1727203874.95992: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking connection profiles [028d2410-947f-82dc-c122-000000000a3e] 13731 1727203874.95994: sending task result for task 028d2410-947f-82dc-c122-000000000a3e 13731 1727203874.96071: done sending task result for task 028d2410-947f-82dc-c122-000000000a3e 13731 1727203874.96074: WORKER PROCESS EXITING changed: [managed-node3] => { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "bond": { "arp_interval": 60, "arp_ip_target": "192.0.2.128", "arp_validate": "none", "mode": "active-backup", "primary": "test1" }, "interface_name": "nm-bond", "ip": { "route_metric4": 65535 }, "name": "bond0", "state": "up", "type": "bond" }, { "controller": "bond0", "interface_name": "test1", "name": "bond0.0", "state": "up", "type": "ethernet" }, { "controller": "bond0", "interface_name": "test2", "name": "bond0.1", "state": "up", "type": "ethernet" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true } STDERR: [007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df [008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 [009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 [010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df (is-modified) [011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 (not-active) [012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 (not-active) 13731 1727203874.96311: no more pending results, returning what we have 13731 1727203874.96314: results queue empty 13731 1727203874.96315: checking for any_errors_fatal 13731 1727203874.96320: done checking for any_errors_fatal 13731 1727203874.96321: checking for max_fail_percentage 13731 1727203874.96323: done checking for max_fail_percentage 13731 1727203874.96323: checking to see if all hosts have failed and the running result is not ok 13731 1727203874.96324: done checking to see if all hosts have failed 13731 1727203874.96325: getting the remaining hosts for this loop 13731 1727203874.96326: done getting the remaining hosts for this loop 13731 1727203874.96329: getting the next task for host managed-node3 13731 1727203874.96336: done getting next task for host managed-node3 13731 1727203874.96339: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking state 13731 1727203874.96343: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=21, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203874.96352: getting variables 13731 1727203874.96354: in VariableManager get_vars() 13731 1727203874.96507: Calling all_inventory to load vars for managed-node3 13731 1727203874.96510: Calling groups_inventory to load vars for managed-node3 13731 1727203874.96512: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203874.96520: Calling all_plugins_play to load vars for managed-node3 13731 1727203874.96522: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203874.96525: Calling groups_plugins_play to load vars for managed-node3 13731 1727203874.98061: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203874.99709: done with get_vars() 13731 1727203874.99742: done getting variables TASK [fedora.linux_system_roles.network : Configure networking state] ********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:171 Tuesday 24 September 2024 14:51:14 -0400 (0:00:00.817) 0:00:53.229 ***** 13731 1727203874.99828: entering _queue_task() for managed-node3/fedora.linux_system_roles.network_state 13731 1727203875.00257: worker is 1 (out of 1 available) 13731 1727203875.00390: exiting _queue_task() for managed-node3/fedora.linux_system_roles.network_state 13731 1727203875.00403: done queuing things up, now waiting for results queue to drain 13731 1727203875.00405: waiting for pending results... 13731 1727203875.00630: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking state 13731 1727203875.00811: in run() - task 028d2410-947f-82dc-c122-000000000a3f 13731 1727203875.00832: variable 'ansible_search_path' from source: unknown 13731 1727203875.00836: variable 'ansible_search_path' from source: unknown 13731 1727203875.00880: calling self._execute() 13731 1727203875.01045: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.01049: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.01053: variable 'omit' from source: magic vars 13731 1727203875.01682: variable 'ansible_distribution_major_version' from source: facts 13731 1727203875.01685: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203875.01688: variable 'network_state' from source: role '' defaults 13731 1727203875.01691: Evaluated conditional (network_state != {}): False 13731 1727203875.01693: when evaluation is False, skipping this task 13731 1727203875.01880: _execute() done 13731 1727203875.01883: dumping result to json 13731 1727203875.01885: done dumping result, returning 13731 1727203875.01887: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking state [028d2410-947f-82dc-c122-000000000a3f] 13731 1727203875.01890: sending task result for task 028d2410-947f-82dc-c122-000000000a3f 13731 1727203875.01952: done sending task result for task 028d2410-947f-82dc-c122-000000000a3f 13731 1727203875.01956: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203875.02019: no more pending results, returning what we have 13731 1727203875.02023: results queue empty 13731 1727203875.02024: checking for any_errors_fatal 13731 1727203875.02035: done checking for any_errors_fatal 13731 1727203875.02036: checking for max_fail_percentage 13731 1727203875.02038: done checking for max_fail_percentage 13731 1727203875.02039: checking to see if all hosts have failed and the running result is not ok 13731 1727203875.02040: done checking to see if all hosts have failed 13731 1727203875.02040: getting the remaining hosts for this loop 13731 1727203875.02042: done getting the remaining hosts for this loop 13731 1727203875.02045: getting the next task for host managed-node3 13731 1727203875.02052: done getting next task for host managed-node3 13731 1727203875.02056: ^ task is: TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13731 1727203875.02061: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=22, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203875.02081: getting variables 13731 1727203875.02083: in VariableManager get_vars() 13731 1727203875.02119: Calling all_inventory to load vars for managed-node3 13731 1727203875.02121: Calling groups_inventory to load vars for managed-node3 13731 1727203875.02124: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203875.02132: Calling all_plugins_play to load vars for managed-node3 13731 1727203875.02135: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203875.02138: Calling groups_plugins_play to load vars for managed-node3 13731 1727203875.03723: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203875.05762: done with get_vars() 13731 1727203875.05787: done getting variables 13731 1727203875.05866: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show stderr messages for the network_connections] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:177 Tuesday 24 September 2024 14:51:15 -0400 (0:00:00.060) 0:00:53.290 ***** 13731 1727203875.05907: entering _queue_task() for managed-node3/debug 13731 1727203875.06485: worker is 1 (out of 1 available) 13731 1727203875.06497: exiting _queue_task() for managed-node3/debug 13731 1727203875.06506: done queuing things up, now waiting for results queue to drain 13731 1727203875.06507: waiting for pending results... 13731 1727203875.06617: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13731 1727203875.06883: in run() - task 028d2410-947f-82dc-c122-000000000a40 13731 1727203875.06887: variable 'ansible_search_path' from source: unknown 13731 1727203875.06890: variable 'ansible_search_path' from source: unknown 13731 1727203875.06892: calling self._execute() 13731 1727203875.06943: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.06954: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.06967: variable 'omit' from source: magic vars 13731 1727203875.07482: variable 'ansible_distribution_major_version' from source: facts 13731 1727203875.07485: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203875.07488: variable 'omit' from source: magic vars 13731 1727203875.07492: variable 'omit' from source: magic vars 13731 1727203875.07510: variable 'omit' from source: magic vars 13731 1727203875.07552: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203875.07592: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203875.07618: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203875.07636: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203875.07649: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203875.07685: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203875.07688: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.07691: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.07982: Set connection var ansible_pipelining to False 13731 1727203875.07985: Set connection var ansible_shell_type to sh 13731 1727203875.07988: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203875.07990: Set connection var ansible_connection to ssh 13731 1727203875.07992: Set connection var ansible_shell_executable to /bin/sh 13731 1727203875.07995: Set connection var ansible_timeout to 10 13731 1727203875.07997: variable 'ansible_shell_executable' from source: unknown 13731 1727203875.07999: variable 'ansible_connection' from source: unknown 13731 1727203875.08002: variable 'ansible_module_compression' from source: unknown 13731 1727203875.08004: variable 'ansible_shell_type' from source: unknown 13731 1727203875.08006: variable 'ansible_shell_executable' from source: unknown 13731 1727203875.08008: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.08010: variable 'ansible_pipelining' from source: unknown 13731 1727203875.08012: variable 'ansible_timeout' from source: unknown 13731 1727203875.08014: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.08027: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203875.08045: variable 'omit' from source: magic vars 13731 1727203875.08052: starting attempt loop 13731 1727203875.08054: running the handler 13731 1727203875.08223: variable '__network_connections_result' from source: set_fact 13731 1727203875.08402: handler run complete 13731 1727203875.08418: attempt loop complete, returning result 13731 1727203875.08422: _execute() done 13731 1727203875.08424: dumping result to json 13731 1727203875.08427: done dumping result, returning 13731 1727203875.08435: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections [028d2410-947f-82dc-c122-000000000a40] 13731 1727203875.08441: sending task result for task 028d2410-947f-82dc-c122-000000000a40 ok: [managed-node3] => { "__network_connections_result.stderr_lines": [ "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df", "[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999", "[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82", "[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df (is-modified)", "[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 (not-active)", "[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 (not-active)" ] } 13731 1727203875.08604: done sending task result for task 028d2410-947f-82dc-c122-000000000a40 13731 1727203875.08608: WORKER PROCESS EXITING 13731 1727203875.08836: no more pending results, returning what we have 13731 1727203875.08841: results queue empty 13731 1727203875.08841: checking for any_errors_fatal 13731 1727203875.08849: done checking for any_errors_fatal 13731 1727203875.08850: checking for max_fail_percentage 13731 1727203875.08852: done checking for max_fail_percentage 13731 1727203875.08853: checking to see if all hosts have failed and the running result is not ok 13731 1727203875.08853: done checking to see if all hosts have failed 13731 1727203875.08854: getting the remaining hosts for this loop 13731 1727203875.08856: done getting the remaining hosts for this loop 13731 1727203875.08861: getting the next task for host managed-node3 13731 1727203875.08869: done getting next task for host managed-node3 13731 1727203875.08873: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13731 1727203875.08879: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=23, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203875.08891: getting variables 13731 1727203875.08892: in VariableManager get_vars() 13731 1727203875.09281: Calling all_inventory to load vars for managed-node3 13731 1727203875.09285: Calling groups_inventory to load vars for managed-node3 13731 1727203875.09287: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203875.09297: Calling all_plugins_play to load vars for managed-node3 13731 1727203875.09306: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203875.09309: Calling groups_plugins_play to load vars for managed-node3 13731 1727203875.17567: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203875.19339: done with get_vars() 13731 1727203875.19371: done getting variables 13731 1727203875.19431: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_connections] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:181 Tuesday 24 September 2024 14:51:15 -0400 (0:00:00.135) 0:00:53.425 ***** 13731 1727203875.19468: entering _queue_task() for managed-node3/debug 13731 1727203875.19839: worker is 1 (out of 1 available) 13731 1727203875.20082: exiting _queue_task() for managed-node3/debug 13731 1727203875.20092: done queuing things up, now waiting for results queue to drain 13731 1727203875.20094: waiting for pending results... 13731 1727203875.20173: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13731 1727203875.20482: in run() - task 028d2410-947f-82dc-c122-000000000a41 13731 1727203875.20486: variable 'ansible_search_path' from source: unknown 13731 1727203875.20490: variable 'ansible_search_path' from source: unknown 13731 1727203875.20494: calling self._execute() 13731 1727203875.20504: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.20513: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.20523: variable 'omit' from source: magic vars 13731 1727203875.20922: variable 'ansible_distribution_major_version' from source: facts 13731 1727203875.20934: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203875.20939: variable 'omit' from source: magic vars 13731 1727203875.21281: variable 'omit' from source: magic vars 13731 1727203875.21284: variable 'omit' from source: magic vars 13731 1727203875.21287: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203875.21290: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203875.21293: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203875.21299: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203875.21302: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203875.21304: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203875.21307: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.21309: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.21335: Set connection var ansible_pipelining to False 13731 1727203875.21338: Set connection var ansible_shell_type to sh 13731 1727203875.21345: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203875.21350: Set connection var ansible_connection to ssh 13731 1727203875.21356: Set connection var ansible_shell_executable to /bin/sh 13731 1727203875.21364: Set connection var ansible_timeout to 10 13731 1727203875.21389: variable 'ansible_shell_executable' from source: unknown 13731 1727203875.21392: variable 'ansible_connection' from source: unknown 13731 1727203875.21395: variable 'ansible_module_compression' from source: unknown 13731 1727203875.21397: variable 'ansible_shell_type' from source: unknown 13731 1727203875.21400: variable 'ansible_shell_executable' from source: unknown 13731 1727203875.21402: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.21411: variable 'ansible_pipelining' from source: unknown 13731 1727203875.21414: variable 'ansible_timeout' from source: unknown 13731 1727203875.21418: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.21782: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203875.21786: variable 'omit' from source: magic vars 13731 1727203875.21788: starting attempt loop 13731 1727203875.21790: running the handler 13731 1727203875.21792: variable '__network_connections_result' from source: set_fact 13731 1727203875.21794: variable '__network_connections_result' from source: set_fact 13731 1727203875.21906: handler run complete 13731 1727203875.21935: attempt loop complete, returning result 13731 1727203875.21938: _execute() done 13731 1727203875.21941: dumping result to json 13731 1727203875.21946: done dumping result, returning 13731 1727203875.21961: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections [028d2410-947f-82dc-c122-000000000a41] 13731 1727203875.21969: sending task result for task 028d2410-947f-82dc-c122-000000000a41 ok: [managed-node3] => { "__network_connections_result": { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "bond": { "arp_interval": 60, "arp_ip_target": "192.0.2.128", "arp_validate": "none", "mode": "active-backup", "primary": "test1" }, "interface_name": "nm-bond", "ip": { "route_metric4": 65535 }, "name": "bond0", "state": "up", "type": "bond" }, { "controller": "bond0", "interface_name": "test1", "name": "bond0.0", "state": "up", "type": "ethernet" }, { "controller": "bond0", "interface_name": "test2", "name": "bond0.1", "state": "up", "type": "ethernet" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true, "failed": false, "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 (not-active)\n", "stderr_lines": [ "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df", "[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999", "[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82", "[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df (is-modified)", "[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 (not-active)", "[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 (not-active)" ] } } 13731 1727203875.22203: no more pending results, returning what we have 13731 1727203875.22208: results queue empty 13731 1727203875.22215: checking for any_errors_fatal 13731 1727203875.22225: done checking for any_errors_fatal 13731 1727203875.22226: checking for max_fail_percentage 13731 1727203875.22228: done checking for max_fail_percentage 13731 1727203875.22229: checking to see if all hosts have failed and the running result is not ok 13731 1727203875.22230: done checking to see if all hosts have failed 13731 1727203875.22230: getting the remaining hosts for this loop 13731 1727203875.22232: done getting the remaining hosts for this loop 13731 1727203875.22235: getting the next task for host managed-node3 13731 1727203875.22243: done getting next task for host managed-node3 13731 1727203875.22247: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13731 1727203875.22254: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=24, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203875.22270: getting variables 13731 1727203875.22272: in VariableManager get_vars() 13731 1727203875.22428: Calling all_inventory to load vars for managed-node3 13731 1727203875.22432: Calling groups_inventory to load vars for managed-node3 13731 1727203875.22434: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203875.22445: Calling all_plugins_play to load vars for managed-node3 13731 1727203875.22448: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203875.22452: Calling groups_plugins_play to load vars for managed-node3 13731 1727203875.22973: done sending task result for task 028d2410-947f-82dc-c122-000000000a41 13731 1727203875.22979: WORKER PROCESS EXITING 13731 1727203875.24352: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203875.26223: done with get_vars() 13731 1727203875.26252: done getting variables 13731 1727203875.26340: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_state] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:186 Tuesday 24 September 2024 14:51:15 -0400 (0:00:00.069) 0:00:53.494 ***** 13731 1727203875.26405: entering _queue_task() for managed-node3/debug 13731 1727203875.26896: worker is 1 (out of 1 available) 13731 1727203875.26916: exiting _queue_task() for managed-node3/debug 13731 1727203875.26928: done queuing things up, now waiting for results queue to drain 13731 1727203875.26930: waiting for pending results... 13731 1727203875.27277: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13731 1727203875.27434: in run() - task 028d2410-947f-82dc-c122-000000000a42 13731 1727203875.27456: variable 'ansible_search_path' from source: unknown 13731 1727203875.27465: variable 'ansible_search_path' from source: unknown 13731 1727203875.27505: calling self._execute() 13731 1727203875.27601: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.27612: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.27623: variable 'omit' from source: magic vars 13731 1727203875.27990: variable 'ansible_distribution_major_version' from source: facts 13731 1727203875.28006: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203875.28123: variable 'network_state' from source: role '' defaults 13731 1727203875.28139: Evaluated conditional (network_state != {}): False 13731 1727203875.28146: when evaluation is False, skipping this task 13731 1727203875.28153: _execute() done 13731 1727203875.28160: dumping result to json 13731 1727203875.28167: done dumping result, returning 13731 1727203875.28179: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state [028d2410-947f-82dc-c122-000000000a42] 13731 1727203875.28189: sending task result for task 028d2410-947f-82dc-c122-000000000a42 skipping: [managed-node3] => { "false_condition": "network_state != {}" } 13731 1727203875.28657: no more pending results, returning what we have 13731 1727203875.28664: results queue empty 13731 1727203875.28664: checking for any_errors_fatal 13731 1727203875.28781: done checking for any_errors_fatal 13731 1727203875.28782: checking for max_fail_percentage 13731 1727203875.28786: done checking for max_fail_percentage 13731 1727203875.28787: checking to see if all hosts have failed and the running result is not ok 13731 1727203875.28788: done checking to see if all hosts have failed 13731 1727203875.28788: getting the remaining hosts for this loop 13731 1727203875.28791: done getting the remaining hosts for this loop 13731 1727203875.28794: getting the next task for host managed-node3 13731 1727203875.28802: done getting next task for host managed-node3 13731 1727203875.28806: ^ task is: TASK: fedora.linux_system_roles.network : Re-test connectivity 13731 1727203875.28812: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=25, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203875.28830: done sending task result for task 028d2410-947f-82dc-c122-000000000a42 13731 1727203875.28834: WORKER PROCESS EXITING 13731 1727203875.28848: getting variables 13731 1727203875.28850: in VariableManager get_vars() 13731 1727203875.28897: Calling all_inventory to load vars for managed-node3 13731 1727203875.28900: Calling groups_inventory to load vars for managed-node3 13731 1727203875.28903: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203875.28915: Calling all_plugins_play to load vars for managed-node3 13731 1727203875.28918: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203875.28921: Calling groups_plugins_play to load vars for managed-node3 13731 1727203875.31718: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203875.34784: done with get_vars() 13731 1727203875.34815: done getting variables TASK [fedora.linux_system_roles.network : Re-test connectivity] **************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:192 Tuesday 24 September 2024 14:51:15 -0400 (0:00:00.085) 0:00:53.580 ***** 13731 1727203875.34933: entering _queue_task() for managed-node3/ping 13731 1727203875.35493: worker is 1 (out of 1 available) 13731 1727203875.35503: exiting _queue_task() for managed-node3/ping 13731 1727203875.35513: done queuing things up, now waiting for results queue to drain 13731 1727203875.35515: waiting for pending results... 13731 1727203875.35664: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Re-test connectivity 13731 1727203875.36082: in run() - task 028d2410-947f-82dc-c122-000000000a43 13731 1727203875.36086: variable 'ansible_search_path' from source: unknown 13731 1727203875.36089: variable 'ansible_search_path' from source: unknown 13731 1727203875.36093: calling self._execute() 13731 1727203875.36096: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.36098: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.36100: variable 'omit' from source: magic vars 13731 1727203875.36432: variable 'ansible_distribution_major_version' from source: facts 13731 1727203875.36470: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203875.36473: variable 'omit' from source: magic vars 13731 1727203875.36528: variable 'omit' from source: magic vars 13731 1727203875.36567: variable 'omit' from source: magic vars 13731 1727203875.36616: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203875.36657: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203875.36674: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203875.36694: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203875.36706: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203875.36784: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203875.36787: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.36790: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.36941: Set connection var ansible_pipelining to False 13731 1727203875.36944: Set connection var ansible_shell_type to sh 13731 1727203875.36946: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203875.36948: Set connection var ansible_connection to ssh 13731 1727203875.36950: Set connection var ansible_shell_executable to /bin/sh 13731 1727203875.36952: Set connection var ansible_timeout to 10 13731 1727203875.36954: variable 'ansible_shell_executable' from source: unknown 13731 1727203875.36956: variable 'ansible_connection' from source: unknown 13731 1727203875.36958: variable 'ansible_module_compression' from source: unknown 13731 1727203875.36960: variable 'ansible_shell_type' from source: unknown 13731 1727203875.36961: variable 'ansible_shell_executable' from source: unknown 13731 1727203875.36963: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.36965: variable 'ansible_pipelining' from source: unknown 13731 1727203875.36968: variable 'ansible_timeout' from source: unknown 13731 1727203875.36970: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.37156: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203875.37181: variable 'omit' from source: magic vars 13731 1727203875.37184: starting attempt loop 13731 1727203875.37186: running the handler 13731 1727203875.37380: _low_level_execute_command(): starting 13731 1727203875.37384: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203875.37996: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203875.38053: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203875.38078: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203875.38145: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203875.38247: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203875.39910: stdout chunk (state=3): >>>/root <<< 13731 1727203875.40045: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203875.40049: stdout chunk (state=3): >>><<< 13731 1727203875.40058: stderr chunk (state=3): >>><<< 13731 1727203875.40202: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203875.40282: _low_level_execute_command(): starting 13731 1727203875.40286: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635 `" && echo ansible-tmp-1727203875.402021-17837-145158767841635="` echo /root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635 `" ) && sleep 0' 13731 1727203875.41289: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203875.41460: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203875.41464: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203875.41468: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203875.41481: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203875.41484: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203875.41612: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203875.41688: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203875.43704: stdout chunk (state=3): >>>ansible-tmp-1727203875.402021-17837-145158767841635=/root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635 <<< 13731 1727203875.43726: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203875.43773: stderr chunk (state=3): >>><<< 13731 1727203875.43779: stdout chunk (state=3): >>><<< 13731 1727203875.43805: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203875.402021-17837-145158767841635=/root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203875.43853: variable 'ansible_module_compression' from source: unknown 13731 1727203875.44129: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.ping-ZIP_DEFLATED 13731 1727203875.44166: variable 'ansible_facts' from source: unknown 13731 1727203875.44458: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/AnsiballZ_ping.py 13731 1727203875.45082: Sending initial data 13731 1727203875.45085: Sent initial data (152 bytes) 13731 1727203875.46059: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203875.46072: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203875.46085: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203875.46099: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203875.46112: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203875.46118: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203875.46230: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203875.46294: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203875.46474: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203875.47936: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203875.47968: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203875.48005: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpngdxuolg /root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/AnsiballZ_ping.py <<< 13731 1727203875.48009: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/AnsiballZ_ping.py" <<< 13731 1727203875.48042: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpngdxuolg" to remote "/root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/AnsiballZ_ping.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/AnsiballZ_ping.py" <<< 13731 1727203875.49257: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203875.49260: stdout chunk (state=3): >>><<< 13731 1727203875.49372: stderr chunk (state=3): >>><<< 13731 1727203875.49527: done transferring module to remote 13731 1727203875.49538: _low_level_execute_command(): starting 13731 1727203875.49543: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/ /root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/AnsiballZ_ping.py && sleep 0' 13731 1727203875.50601: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203875.50741: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203875.50759: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203875.50780: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203875.50820: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203875.50916: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203875.51290: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203875.51304: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203875.51377: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203875.53472: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203875.53485: stdout chunk (state=3): >>><<< 13731 1727203875.53497: stderr chunk (state=3): >>><<< 13731 1727203875.53516: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203875.53525: _low_level_execute_command(): starting 13731 1727203875.53535: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/AnsiballZ_ping.py && sleep 0' 13731 1727203875.54716: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203875.54731: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203875.54744: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203875.54914: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203875.54926: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203875.55007: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203875.55215: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203875.70344: stdout chunk (state=3): >>> {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} <<< 13731 1727203875.71621: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203875.71625: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203875.71685: stderr chunk (state=3): >>><<< 13731 1727203875.71688: stdout chunk (state=3): >>><<< 13731 1727203875.71700: _low_level_execute_command() done: rc=0, stdout= {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203875.71725: done with _execute_module (ping, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ping', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203875.71803: _low_level_execute_command(): starting 13731 1727203875.71806: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203875.402021-17837-145158767841635/ > /dev/null 2>&1 && sleep 0' 13731 1727203875.72313: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203875.72321: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203875.72332: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203875.72346: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203875.72361: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203875.72369: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203875.72381: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203875.72397: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203875.72531: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203875.72535: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203875.72537: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203875.72539: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203875.72541: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203875.72543: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203875.72545: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203875.72546: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203875.72553: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203875.72556: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203875.72590: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203875.72663: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203875.74638: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203875.74643: stdout chunk (state=3): >>><<< 13731 1727203875.74645: stderr chunk (state=3): >>><<< 13731 1727203875.74666: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203875.74748: handler run complete 13731 1727203875.74751: attempt loop complete, returning result 13731 1727203875.74753: _execute() done 13731 1727203875.74756: dumping result to json 13731 1727203875.74778: done dumping result, returning 13731 1727203875.74816: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Re-test connectivity [028d2410-947f-82dc-c122-000000000a43] 13731 1727203875.74819: sending task result for task 028d2410-947f-82dc-c122-000000000a43 ok: [managed-node3] => { "changed": false, "ping": "pong" } 13731 1727203875.75241: no more pending results, returning what we have 13731 1727203875.75246: results queue empty 13731 1727203875.75247: checking for any_errors_fatal 13731 1727203875.75253: done checking for any_errors_fatal 13731 1727203875.75254: checking for max_fail_percentage 13731 1727203875.75255: done checking for max_fail_percentage 13731 1727203875.75256: checking to see if all hosts have failed and the running result is not ok 13731 1727203875.75257: done checking to see if all hosts have failed 13731 1727203875.75260: getting the remaining hosts for this loop 13731 1727203875.75262: done getting the remaining hosts for this loop 13731 1727203875.75266: getting the next task for host managed-node3 13731 1727203875.75300: done getting next task for host managed-node3 13731 1727203875.75303: ^ task is: TASK: meta (role_complete) 13731 1727203875.75380: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203875.75402: getting variables 13731 1727203875.75404: in VariableManager get_vars() 13731 1727203875.75450: Calling all_inventory to load vars for managed-node3 13731 1727203875.75454: Calling groups_inventory to load vars for managed-node3 13731 1727203875.75457: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203875.75470: Calling all_plugins_play to load vars for managed-node3 13731 1727203875.75474: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203875.75940: Calling groups_plugins_play to load vars for managed-node3 13731 1727203875.76474: done sending task result for task 028d2410-947f-82dc-c122-000000000a43 13731 1727203875.76480: WORKER PROCESS EXITING 13731 1727203875.77686: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203875.80108: done with get_vars() 13731 1727203875.80152: done getting variables 13731 1727203875.80267: done queuing things up, now waiting for results queue to drain 13731 1727203875.80270: results queue empty 13731 1727203875.80270: checking for any_errors_fatal 13731 1727203875.80273: done checking for any_errors_fatal 13731 1727203875.80274: checking for max_fail_percentage 13731 1727203875.80278: done checking for max_fail_percentage 13731 1727203875.80279: checking to see if all hosts have failed and the running result is not ok 13731 1727203875.80280: done checking to see if all hosts have failed 13731 1727203875.80281: getting the remaining hosts for this loop 13731 1727203875.80282: done getting the remaining hosts for this loop 13731 1727203875.80285: getting the next task for host managed-node3 13731 1727203875.80290: done getting next task for host managed-node3 13731 1727203875.80293: ^ task is: TASK: Show result 13731 1727203875.80296: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203875.80299: getting variables 13731 1727203875.80300: in VariableManager get_vars() 13731 1727203875.80317: Calling all_inventory to load vars for managed-node3 13731 1727203875.80319: Calling groups_inventory to load vars for managed-node3 13731 1727203875.80321: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203875.80327: Calling all_plugins_play to load vars for managed-node3 13731 1727203875.80330: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203875.80332: Calling groups_plugins_play to load vars for managed-node3 13731 1727203875.81565: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203875.84193: done with get_vars() 13731 1727203875.84224: done getting variables 13731 1727203875.84279: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Show result] ************************************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_bond_profile_reconfigure.yml:33 Tuesday 24 September 2024 14:51:15 -0400 (0:00:00.493) 0:00:54.074 ***** 13731 1727203875.84317: entering _queue_task() for managed-node3/debug 13731 1727203875.84806: worker is 1 (out of 1 available) 13731 1727203875.84817: exiting _queue_task() for managed-node3/debug 13731 1727203875.84827: done queuing things up, now waiting for results queue to drain 13731 1727203875.84828: waiting for pending results... 13731 1727203875.85028: running TaskExecutor() for managed-node3/TASK: Show result 13731 1727203875.85162: in run() - task 028d2410-947f-82dc-c122-000000000a73 13731 1727203875.85171: variable 'ansible_search_path' from source: unknown 13731 1727203875.85234: variable 'ansible_search_path' from source: unknown 13731 1727203875.85237: calling self._execute() 13731 1727203875.85327: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.85345: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.85362: variable 'omit' from source: magic vars 13731 1727203875.85786: variable 'ansible_distribution_major_version' from source: facts 13731 1727203875.85805: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203875.85819: variable 'omit' from source: magic vars 13731 1727203875.85883: variable 'omit' from source: magic vars 13731 1727203875.85887: variable 'omit' from source: magic vars 13731 1727203875.85938: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203875.85979: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203875.86009: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203875.86035: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203875.86055: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203875.86091: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203875.86180: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.86184: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.86226: Set connection var ansible_pipelining to False 13731 1727203875.86236: Set connection var ansible_shell_type to sh 13731 1727203875.86245: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203875.86255: Set connection var ansible_connection to ssh 13731 1727203875.86264: Set connection var ansible_shell_executable to /bin/sh 13731 1727203875.86273: Set connection var ansible_timeout to 10 13731 1727203875.86304: variable 'ansible_shell_executable' from source: unknown 13731 1727203875.86316: variable 'ansible_connection' from source: unknown 13731 1727203875.86325: variable 'ansible_module_compression' from source: unknown 13731 1727203875.86408: variable 'ansible_shell_type' from source: unknown 13731 1727203875.86411: variable 'ansible_shell_executable' from source: unknown 13731 1727203875.86413: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.86415: variable 'ansible_pipelining' from source: unknown 13731 1727203875.86417: variable 'ansible_timeout' from source: unknown 13731 1727203875.86423: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.86503: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203875.86525: variable 'omit' from source: magic vars 13731 1727203875.86538: starting attempt loop 13731 1727203875.86544: running the handler 13731 1727203875.86594: variable '__network_connections_result' from source: set_fact 13731 1727203875.86684: variable '__network_connections_result' from source: set_fact 13731 1727203875.86888: handler run complete 13731 1727203875.86923: attempt loop complete, returning result 13731 1727203875.86930: _execute() done 13731 1727203875.86935: dumping result to json 13731 1727203875.86948: done dumping result, returning 13731 1727203875.86959: done running TaskExecutor() for managed-node3/TASK: Show result [028d2410-947f-82dc-c122-000000000a73] 13731 1727203875.86973: sending task result for task 028d2410-947f-82dc-c122-000000000a73 ok: [managed-node3] => { "__network_connections_result": { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "bond": { "arp_interval": 60, "arp_ip_target": "192.0.2.128", "arp_validate": "none", "mode": "active-backup", "primary": "test1" }, "interface_name": "nm-bond", "ip": { "route_metric4": 65535 }, "name": "bond0", "state": "up", "type": "bond" }, { "controller": "bond0", "interface_name": "test1", "name": "bond0.0", "state": "up", "type": "ethernet" }, { "controller": "bond0", "interface_name": "test2", "name": "bond0.1", "state": "up", "type": "ethernet" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true, "failed": false, "stderr": "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df\n[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999\n[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82\n[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df (is-modified)\n[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 (not-active)\n[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 (not-active)\n", "stderr_lines": [ "[007] #0, state:up persistent_state:present, 'bond0': add connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df", "[008] #1, state:up persistent_state:present, 'bond0.0': add connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999", "[009] #2, state:up persistent_state:present, 'bond0.1': add connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82", "[010] #0, state:up persistent_state:present, 'bond0': up connection bond0, 65e7c30c-f406-45a4-954f-9e2a9480d1df (is-modified)", "[011] #1, state:up persistent_state:present, 'bond0.0': up connection bond0.0, 76f9383b-e1a2-4403-937c-02353034d999 (not-active)", "[012] #2, state:up persistent_state:present, 'bond0.1': up connection bond0.1, 9eb2a8f6-720f-41da-85c9-7f6979308b82 (not-active)" ] } } 13731 1727203875.87306: no more pending results, returning what we have 13731 1727203875.87312: results queue empty 13731 1727203875.87313: checking for any_errors_fatal 13731 1727203875.87315: done checking for any_errors_fatal 13731 1727203875.87385: checking for max_fail_percentage 13731 1727203875.87389: done checking for max_fail_percentage 13731 1727203875.87390: checking to see if all hosts have failed and the running result is not ok 13731 1727203875.87391: done checking to see if all hosts have failed 13731 1727203875.87391: getting the remaining hosts for this loop 13731 1727203875.87397: done getting the remaining hosts for this loop 13731 1727203875.87401: getting the next task for host managed-node3 13731 1727203875.87413: done getting next task for host managed-node3 13731 1727203875.87417: ^ task is: TASK: Asserts 13731 1727203875.87420: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203875.87426: getting variables 13731 1727203875.87428: in VariableManager get_vars() 13731 1727203875.87472: Calling all_inventory to load vars for managed-node3 13731 1727203875.87323: done sending task result for task 028d2410-947f-82dc-c122-000000000a73 13731 1727203875.87624: WORKER PROCESS EXITING 13731 1727203875.87618: Calling groups_inventory to load vars for managed-node3 13731 1727203875.87629: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203875.87639: Calling all_plugins_play to load vars for managed-node3 13731 1727203875.87642: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203875.87646: Calling groups_plugins_play to load vars for managed-node3 13731 1727203875.89195: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203875.90774: done with get_vars() 13731 1727203875.90806: done getting variables TASK [Asserts] ***************************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:36 Tuesday 24 September 2024 14:51:15 -0400 (0:00:00.065) 0:00:54.139 ***** 13731 1727203875.90900: entering _queue_task() for managed-node3/include_tasks 13731 1727203875.91251: worker is 1 (out of 1 available) 13731 1727203875.91263: exiting _queue_task() for managed-node3/include_tasks 13731 1727203875.91481: done queuing things up, now waiting for results queue to drain 13731 1727203875.91483: waiting for pending results... 13731 1727203875.91563: running TaskExecutor() for managed-node3/TASK: Asserts 13731 1727203875.91703: in run() - task 028d2410-947f-82dc-c122-0000000008ef 13731 1727203875.91727: variable 'ansible_search_path' from source: unknown 13731 1727203875.91735: variable 'ansible_search_path' from source: unknown 13731 1727203875.91789: variable 'lsr_assert' from source: include params 13731 1727203875.92036: variable 'lsr_assert' from source: include params 13731 1727203875.92112: variable 'omit' from source: magic vars 13731 1727203875.92268: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203875.92286: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203875.92303: variable 'omit' from source: magic vars 13731 1727203875.92559: variable 'ansible_distribution_major_version' from source: facts 13731 1727203875.92582: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203875.92684: variable 'item' from source: unknown 13731 1727203875.92687: variable 'item' from source: unknown 13731 1727203875.92703: variable 'item' from source: unknown 13731 1727203875.92763: variable 'item' from source: unknown 13731 1727203875.92999: dumping result to json 13731 1727203875.93007: done dumping result, returning 13731 1727203875.93010: done running TaskExecutor() for managed-node3/TASK: Asserts [028d2410-947f-82dc-c122-0000000008ef] 13731 1727203875.93012: sending task result for task 028d2410-947f-82dc-c122-0000000008ef 13731 1727203875.93055: done sending task result for task 028d2410-947f-82dc-c122-0000000008ef 13731 1727203875.93060: WORKER PROCESS EXITING 13731 1727203875.93126: no more pending results, returning what we have 13731 1727203875.93132: in VariableManager get_vars() 13731 1727203875.93185: Calling all_inventory to load vars for managed-node3 13731 1727203875.93189: Calling groups_inventory to load vars for managed-node3 13731 1727203875.93192: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203875.93206: Calling all_plugins_play to load vars for managed-node3 13731 1727203875.93209: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203875.93213: Calling groups_plugins_play to load vars for managed-node3 13731 1727203875.94726: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203875.96408: done with get_vars() 13731 1727203875.96427: variable 'ansible_search_path' from source: unknown 13731 1727203875.96429: variable 'ansible_search_path' from source: unknown 13731 1727203875.96469: we have included files to process 13731 1727203875.96470: generating all_blocks data 13731 1727203875.96472: done generating all_blocks data 13731 1727203875.96479: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml 13731 1727203875.96480: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml 13731 1727203875.96482: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml 13731 1727203875.96733: in VariableManager get_vars() 13731 1727203875.96758: done with get_vars() 13731 1727203875.96801: in VariableManager get_vars() 13731 1727203875.96823: done with get_vars() 13731 1727203875.96837: done processing included file 13731 1727203875.96839: iterating over new_blocks loaded from include file 13731 1727203875.96840: in VariableManager get_vars() 13731 1727203875.96857: done with get_vars() 13731 1727203875.96858: filtering new block on tags 13731 1727203875.96902: done filtering new block on tags 13731 1727203875.96905: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml for managed-node3 => (item=tasks/assert_bond_options.yml) 13731 1727203875.96910: extending task lists for all hosts with included blocks 13731 1727203875.99986: done extending task lists 13731 1727203875.99988: done processing included files 13731 1727203875.99989: results queue empty 13731 1727203875.99990: checking for any_errors_fatal 13731 1727203875.99995: done checking for any_errors_fatal 13731 1727203875.99996: checking for max_fail_percentage 13731 1727203875.99997: done checking for max_fail_percentage 13731 1727203875.99998: checking to see if all hosts have failed and the running result is not ok 13731 1727203875.99999: done checking to see if all hosts have failed 13731 1727203876.00000: getting the remaining hosts for this loop 13731 1727203876.00001: done getting the remaining hosts for this loop 13731 1727203876.00004: getting the next task for host managed-node3 13731 1727203876.00008: done getting next task for host managed-node3 13731 1727203876.00010: ^ task is: TASK: ** TEST check bond settings 13731 1727203876.00013: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203876.00016: getting variables 13731 1727203876.00017: in VariableManager get_vars() 13731 1727203876.00035: Calling all_inventory to load vars for managed-node3 13731 1727203876.00037: Calling groups_inventory to load vars for managed-node3 13731 1727203876.00039: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203876.00046: Calling all_plugins_play to load vars for managed-node3 13731 1727203876.00048: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203876.00051: Calling groups_plugins_play to load vars for managed-node3 13731 1727203876.01242: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203876.03036: done with get_vars() 13731 1727203876.03072: done getting variables 13731 1727203876.03123: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [** TEST check bond settings] ********************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml:3 Tuesday 24 September 2024 14:51:16 -0400 (0:00:00.122) 0:00:54.262 ***** 13731 1727203876.03156: entering _queue_task() for managed-node3/command 13731 1727203876.03516: worker is 1 (out of 1 available) 13731 1727203876.03530: exiting _queue_task() for managed-node3/command 13731 1727203876.03543: done queuing things up, now waiting for results queue to drain 13731 1727203876.03544: waiting for pending results... 13731 1727203876.03904: running TaskExecutor() for managed-node3/TASK: ** TEST check bond settings 13731 1727203876.03951: in run() - task 028d2410-947f-82dc-c122-000000000c2a 13731 1727203876.04023: variable 'ansible_search_path' from source: unknown 13731 1727203876.04032: variable 'ansible_search_path' from source: unknown 13731 1727203876.04137: variable 'bond_options_to_assert' from source: set_fact 13731 1727203876.04386: variable 'bond_options_to_assert' from source: set_fact 13731 1727203876.04522: variable 'omit' from source: magic vars 13731 1727203876.04668: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.04686: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.04702: variable 'omit' from source: magic vars 13731 1727203876.04944: variable 'ansible_distribution_major_version' from source: facts 13731 1727203876.04982: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203876.04986: variable 'omit' from source: magic vars 13731 1727203876.05021: variable 'omit' from source: magic vars 13731 1727203876.05223: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203876.08283: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203876.08341: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203876.08385: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203876.08429: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203876.08460: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203876.08621: variable 'controller_device' from source: play vars 13731 1727203876.08626: variable 'bond_opt' from source: unknown 13731 1727203876.08629: variable 'omit' from source: magic vars 13731 1727203876.08632: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203876.08665: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203876.08692: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203876.08712: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203876.08729: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203876.08764: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203876.08774: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.08785: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.08891: Set connection var ansible_pipelining to False 13731 1727203876.08901: Set connection var ansible_shell_type to sh 13731 1727203876.08947: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203876.08950: Set connection var ansible_connection to ssh 13731 1727203876.08952: Set connection var ansible_shell_executable to /bin/sh 13731 1727203876.08954: Set connection var ansible_timeout to 10 13731 1727203876.08959: variable 'ansible_shell_executable' from source: unknown 13731 1727203876.08966: variable 'ansible_connection' from source: unknown 13731 1727203876.08972: variable 'ansible_module_compression' from source: unknown 13731 1727203876.08981: variable 'ansible_shell_type' from source: unknown 13731 1727203876.08988: variable 'ansible_shell_executable' from source: unknown 13731 1727203876.08995: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.09002: variable 'ansible_pipelining' from source: unknown 13731 1727203876.09057: variable 'ansible_timeout' from source: unknown 13731 1727203876.09060: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.09130: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203876.09145: variable 'omit' from source: magic vars 13731 1727203876.09155: starting attempt loop 13731 1727203876.09168: running the handler 13731 1727203876.09192: _low_level_execute_command(): starting 13731 1727203876.09203: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203876.09895: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.09935: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.10034: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203876.10039: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.10060: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.10090: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.10160: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.11926: stdout chunk (state=3): >>>/root <<< 13731 1727203876.12046: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.12399: stderr chunk (state=3): >>><<< 13731 1727203876.12402: stdout chunk (state=3): >>><<< 13731 1727203876.12406: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.12417: _low_level_execute_command(): starting 13731 1727203876.12420: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646 `" && echo ansible-tmp-1727203876.1231568-17865-13980375529646="` echo /root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646 `" ) && sleep 0' 13731 1727203876.13596: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.13770: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.13796: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.13871: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.15736: stdout chunk (state=3): >>>ansible-tmp-1727203876.1231568-17865-13980375529646=/root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646 <<< 13731 1727203876.15902: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.15905: stdout chunk (state=3): >>><<< 13731 1727203876.15908: stderr chunk (state=3): >>><<< 13731 1727203876.16086: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203876.1231568-17865-13980375529646=/root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.16090: variable 'ansible_module_compression' from source: unknown 13731 1727203876.16092: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203876.16094: variable 'ansible_facts' from source: unknown 13731 1727203876.16234: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/AnsiballZ_command.py 13731 1727203876.16891: Sending initial data 13731 1727203876.16894: Sent initial data (155 bytes) 13731 1727203876.18277: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.18401: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.18442: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.18474: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.20107: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203876.20238: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203876.20268: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp70koj03z /root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/AnsiballZ_command.py <<< 13731 1727203876.20282: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/AnsiballZ_command.py" <<< 13731 1727203876.20559: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp70koj03z" to remote "/root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/AnsiballZ_command.py" <<< 13731 1727203876.22163: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.22199: stderr chunk (state=3): >>><<< 13731 1727203876.22208: stdout chunk (state=3): >>><<< 13731 1727203876.22348: done transferring module to remote 13731 1727203876.22368: _low_level_execute_command(): starting 13731 1727203876.22408: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/ /root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/AnsiballZ_command.py && sleep 0' 13731 1727203876.23815: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.23883: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.23903: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.23923: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.24154: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.26288: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.26292: stdout chunk (state=3): >>><<< 13731 1727203876.26294: stderr chunk (state=3): >>><<< 13731 1727203876.26297: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.26299: _low_level_execute_command(): starting 13731 1727203876.26340: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/AnsiballZ_command.py && sleep 0' 13731 1727203876.27273: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.27400: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.27445: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.43130: stdout chunk (state=3): >>> {"changed": true, "stdout": "active-backup 1", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/mode"], "start": "2024-09-24 14:51:16.427649", "end": "2024-09-24 14:51:16.430759", "delta": "0:00:00.003110", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/mode", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203876.44617: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203876.44621: stdout chunk (state=3): >>><<< 13731 1727203876.44623: stderr chunk (state=3): >>><<< 13731 1727203876.44690: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "active-backup 1", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/mode"], "start": "2024-09-24 14:51:16.427649", "end": "2024-09-24 14:51:16.430759", "delta": "0:00:00.003110", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/mode", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203876.44721: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/mode', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203876.44736: _low_level_execute_command(): starting 13731 1727203876.44745: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203876.1231568-17865-13980375529646/ > /dev/null 2>&1 && sleep 0' 13731 1727203876.45488: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.45509: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203876.45528: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203876.45552: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203876.45573: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203876.45662: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.45697: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.45715: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.45738: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.45821: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.47655: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.47673: stdout chunk (state=3): >>><<< 13731 1727203876.47688: stderr chunk (state=3): >>><<< 13731 1727203876.47707: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.47723: handler run complete 13731 1727203876.47895: Evaluated conditional (False): False 13731 1727203876.47951: variable 'bond_opt' from source: unknown 13731 1727203876.47970: variable 'result' from source: set_fact 13731 1727203876.47996: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203876.48029: attempt loop complete, returning result 13731 1727203876.48072: variable 'bond_opt' from source: unknown 13731 1727203876.48168: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'mode', 'value': 'active-backup'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "mode", "value": "active-backup" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/mode" ], "delta": "0:00:00.003110", "end": "2024-09-24 14:51:16.430759", "rc": 0, "start": "2024-09-24 14:51:16.427649" } STDOUT: active-backup 1 13731 1727203876.48694: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.48697: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.48699: variable 'omit' from source: magic vars 13731 1727203876.48781: variable 'ansible_distribution_major_version' from source: facts 13731 1727203876.48785: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203876.48787: variable 'omit' from source: magic vars 13731 1727203876.48789: variable 'omit' from source: magic vars 13731 1727203876.48952: variable 'controller_device' from source: play vars 13731 1727203876.48964: variable 'bond_opt' from source: unknown 13731 1727203876.48988: variable 'omit' from source: magic vars 13731 1727203876.49018: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203876.49038: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203876.49091: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203876.49095: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203876.49097: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.49099: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.49193: Set connection var ansible_pipelining to False 13731 1727203876.49204: Set connection var ansible_shell_type to sh 13731 1727203876.49215: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203876.49237: Set connection var ansible_connection to ssh 13731 1727203876.49246: Set connection var ansible_shell_executable to /bin/sh 13731 1727203876.49281: Set connection var ansible_timeout to 10 13731 1727203876.49287: variable 'ansible_shell_executable' from source: unknown 13731 1727203876.49295: variable 'ansible_connection' from source: unknown 13731 1727203876.49301: variable 'ansible_module_compression' from source: unknown 13731 1727203876.49308: variable 'ansible_shell_type' from source: unknown 13731 1727203876.49346: variable 'ansible_shell_executable' from source: unknown 13731 1727203876.49354: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.49356: variable 'ansible_pipelining' from source: unknown 13731 1727203876.49358: variable 'ansible_timeout' from source: unknown 13731 1727203876.49363: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.49463: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203876.49483: variable 'omit' from source: magic vars 13731 1727203876.49493: starting attempt loop 13731 1727203876.49567: running the handler 13731 1727203876.49570: _low_level_execute_command(): starting 13731 1727203876.49573: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203876.50201: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203876.50215: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203876.50233: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.50277: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.50291: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.50361: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.51980: stdout chunk (state=3): >>>/root <<< 13731 1727203876.52084: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.52087: stdout chunk (state=3): >>><<< 13731 1727203876.52090: stderr chunk (state=3): >>><<< 13731 1727203876.52108: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.52202: _low_level_execute_command(): starting 13731 1727203876.52205: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413 `" && echo ansible-tmp-1727203876.5211747-17865-27610383295413="` echo /root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413 `" ) && sleep 0' 13731 1727203876.52770: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.52773: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.52803: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.52850: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.54672: stdout chunk (state=3): >>>ansible-tmp-1727203876.5211747-17865-27610383295413=/root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413 <<< 13731 1727203876.54780: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.54804: stderr chunk (state=3): >>><<< 13731 1727203876.54807: stdout chunk (state=3): >>><<< 13731 1727203876.54834: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203876.5211747-17865-27610383295413=/root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.54856: variable 'ansible_module_compression' from source: unknown 13731 1727203876.54887: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203876.54925: variable 'ansible_facts' from source: unknown 13731 1727203876.55007: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/AnsiballZ_command.py 13731 1727203876.55201: Sending initial data 13731 1727203876.55204: Sent initial data (155 bytes) 13731 1727203876.55752: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.55775: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203876.55870: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.55902: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.55919: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.55939: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.56011: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.57716: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203876.57784: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203876.57837: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp53j7b70j /root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/AnsiballZ_command.py <<< 13731 1727203876.57843: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/AnsiballZ_command.py" <<< 13731 1727203876.57862: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp53j7b70j" to remote "/root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/AnsiballZ_command.py" <<< 13731 1727203876.58630: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.58634: stderr chunk (state=3): >>><<< 13731 1727203876.58639: stdout chunk (state=3): >>><<< 13731 1727203876.58735: done transferring module to remote 13731 1727203876.58739: _low_level_execute_command(): starting 13731 1727203876.58741: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/ /root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/AnsiballZ_command.py && sleep 0' 13731 1727203876.59399: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.59499: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.59540: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.59555: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.59585: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.59655: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.61593: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.61602: stdout chunk (state=3): >>><<< 13731 1727203876.61614: stderr chunk (state=3): >>><<< 13731 1727203876.61831: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.61834: _low_level_execute_command(): starting 13731 1727203876.61837: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/AnsiballZ_command.py && sleep 0' 13731 1727203876.62525: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.62537: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203876.62548: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203876.62565: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203876.62594: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203876.62667: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.62690: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.62709: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.62752: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.78362: stdout chunk (state=3): >>> {"changed": true, "stdout": "60", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/arp_interval"], "start": "2024-09-24 14:51:16.779985", "end": "2024-09-24 14:51:16.783090", "delta": "0:00:00.003105", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/arp_interval", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203876.79958: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203876.79980: stderr chunk (state=3): >>><<< 13731 1727203876.79991: stdout chunk (state=3): >>><<< 13731 1727203876.80282: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "60", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/arp_interval"], "start": "2024-09-24 14:51:16.779985", "end": "2024-09-24 14:51:16.783090", "delta": "0:00:00.003105", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/arp_interval", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203876.80286: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/arp_interval', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203876.80288: _low_level_execute_command(): starting 13731 1727203876.80291: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203876.5211747-17865-27610383295413/ > /dev/null 2>&1 && sleep 0' 13731 1727203876.81272: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.81290: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203876.81303: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203876.81322: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203876.81394: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.81448: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.81474: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.81492: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.81634: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.83496: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.83537: stdout chunk (state=3): >>><<< 13731 1727203876.83540: stderr chunk (state=3): >>><<< 13731 1727203876.83543: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.83545: handler run complete 13731 1727203876.83644: Evaluated conditional (False): False 13731 1727203876.84083: variable 'bond_opt' from source: unknown 13731 1727203876.84086: variable 'result' from source: set_fact 13731 1727203876.84089: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203876.84091: attempt loop complete, returning result 13731 1727203876.84103: variable 'bond_opt' from source: unknown 13731 1727203876.84221: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'arp_interval', 'value': '60'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "arp_interval", "value": "60" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/arp_interval" ], "delta": "0:00:00.003105", "end": "2024-09-24 14:51:16.783090", "rc": 0, "start": "2024-09-24 14:51:16.779985" } STDOUT: 60 13731 1727203876.84531: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.84534: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.84549: variable 'omit' from source: magic vars 13731 1727203876.84753: variable 'ansible_distribution_major_version' from source: facts 13731 1727203876.84763: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203876.84771: variable 'omit' from source: magic vars 13731 1727203876.84791: variable 'omit' from source: magic vars 13731 1727203876.84953: variable 'controller_device' from source: play vars 13731 1727203876.85078: variable 'bond_opt' from source: unknown 13731 1727203876.85081: variable 'omit' from source: magic vars 13731 1727203876.85083: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203876.85085: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203876.85087: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203876.85089: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203876.85091: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.85093: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.85137: Set connection var ansible_pipelining to False 13731 1727203876.85146: Set connection var ansible_shell_type to sh 13731 1727203876.85155: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203876.85163: Set connection var ansible_connection to ssh 13731 1727203876.85171: Set connection var ansible_shell_executable to /bin/sh 13731 1727203876.85187: Set connection var ansible_timeout to 10 13731 1727203876.85208: variable 'ansible_shell_executable' from source: unknown 13731 1727203876.85215: variable 'ansible_connection' from source: unknown 13731 1727203876.85221: variable 'ansible_module_compression' from source: unknown 13731 1727203876.85226: variable 'ansible_shell_type' from source: unknown 13731 1727203876.85232: variable 'ansible_shell_executable' from source: unknown 13731 1727203876.85237: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203876.85244: variable 'ansible_pipelining' from source: unknown 13731 1727203876.85250: variable 'ansible_timeout' from source: unknown 13731 1727203876.85256: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203876.85349: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203876.85361: variable 'omit' from source: magic vars 13731 1727203876.85369: starting attempt loop 13731 1727203876.85376: running the handler 13731 1727203876.85386: _low_level_execute_command(): starting 13731 1727203876.85400: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203876.85996: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.86090: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.86174: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203876.86198: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.86246: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.86331: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.87969: stdout chunk (state=3): >>>/root <<< 13731 1727203876.88147: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.88151: stdout chunk (state=3): >>><<< 13731 1727203876.88154: stderr chunk (state=3): >>><<< 13731 1727203876.88382: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.88390: _low_level_execute_command(): starting 13731 1727203876.88393: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640 `" && echo ansible-tmp-1727203876.8820152-17865-272950727846640="` echo /root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640 `" ) && sleep 0' 13731 1727203876.89257: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.89287: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203876.89299: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203876.89311: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203876.89324: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203876.89330: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203876.89489: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.89493: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203876.89496: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.89499: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.89501: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.91566: stdout chunk (state=3): >>>ansible-tmp-1727203876.8820152-17865-272950727846640=/root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640 <<< 13731 1727203876.91570: stdout chunk (state=3): >>><<< 13731 1727203876.91572: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.91581: stderr chunk (state=3): >>><<< 13731 1727203876.91795: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203876.8820152-17865-272950727846640=/root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.91815: variable 'ansible_module_compression' from source: unknown 13731 1727203876.91854: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203876.91881: variable 'ansible_facts' from source: unknown 13731 1727203876.91938: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/AnsiballZ_command.py 13731 1727203876.92292: Sending initial data 13731 1727203876.92295: Sent initial data (156 bytes) 13731 1727203876.92732: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.92830: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.92847: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.92918: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.94586: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203876.94656: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203876.94660: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp86m1_8r3 /root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/AnsiballZ_command.py <<< 13731 1727203876.94665: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/AnsiballZ_command.py" <<< 13731 1727203876.94720: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp86m1_8r3" to remote "/root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/AnsiballZ_command.py" <<< 13731 1727203876.95358: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.95527: stderr chunk (state=3): >>><<< 13731 1727203876.95530: stdout chunk (state=3): >>><<< 13731 1727203876.95532: done transferring module to remote 13731 1727203876.95535: _low_level_execute_command(): starting 13731 1727203876.95537: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/ /root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/AnsiballZ_command.py && sleep 0' 13731 1727203876.96155: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203876.96189: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203876.96297: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.96317: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.96383: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203876.98150: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203876.98154: stdout chunk (state=3): >>><<< 13731 1727203876.98159: stderr chunk (state=3): >>><<< 13731 1727203876.98246: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203876.98250: _low_level_execute_command(): starting 13731 1727203876.98254: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/AnsiballZ_command.py && sleep 0' 13731 1727203876.98941: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203876.98947: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203876.98965: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203876.98968: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.99046: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203876.99050: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203876.99066: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203876.99151: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203876.99154: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.14917: stdout chunk (state=3): >>> {"changed": true, "stdout": "192.0.2.128", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/arp_ip_target"], "start": "2024-09-24 14:51:17.145107", "end": "2024-09-24 14:51:17.148301", "delta": "0:00:00.003194", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/arp_ip_target", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203877.16487: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203877.16492: stdout chunk (state=3): >>><<< 13731 1727203877.16494: stderr chunk (state=3): >>><<< 13731 1727203877.16497: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "192.0.2.128", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/arp_ip_target"], "start": "2024-09-24 14:51:17.145107", "end": "2024-09-24 14:51:17.148301", "delta": "0:00:00.003194", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/arp_ip_target", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203877.16499: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/arp_ip_target', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203877.16506: _low_level_execute_command(): starting 13731 1727203877.16508: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203876.8820152-17865-272950727846640/ > /dev/null 2>&1 && sleep 0' 13731 1727203877.17156: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203877.17178: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.17196: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.17215: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203877.17256: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.17278: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203877.17368: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.17440: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.17456: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.19282: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.19288: stderr chunk (state=3): >>><<< 13731 1727203877.19291: stdout chunk (state=3): >>><<< 13731 1727203877.19306: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.19309: handler run complete 13731 1727203877.19325: Evaluated conditional (False): False 13731 1727203877.19439: variable 'bond_opt' from source: unknown 13731 1727203877.19445: variable 'result' from source: set_fact 13731 1727203877.19457: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203877.19466: attempt loop complete, returning result 13731 1727203877.19485: variable 'bond_opt' from source: unknown 13731 1727203877.19528: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'arp_ip_target', 'value': '192.0.2.128'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "arp_ip_target", "value": "192.0.2.128" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/arp_ip_target" ], "delta": "0:00:00.003194", "end": "2024-09-24 14:51:17.148301", "rc": 0, "start": "2024-09-24 14:51:17.145107" } STDOUT: 192.0.2.128 13731 1727203877.19654: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203877.19657: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203877.19659: variable 'omit' from source: magic vars 13731 1727203877.19741: variable 'ansible_distribution_major_version' from source: facts 13731 1727203877.19744: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203877.19748: variable 'omit' from source: magic vars 13731 1727203877.19760: variable 'omit' from source: magic vars 13731 1727203877.19868: variable 'controller_device' from source: play vars 13731 1727203877.19871: variable 'bond_opt' from source: unknown 13731 1727203877.19890: variable 'omit' from source: magic vars 13731 1727203877.19904: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203877.19911: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203877.19916: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203877.19926: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203877.19929: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203877.19931: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203877.19979: Set connection var ansible_pipelining to False 13731 1727203877.19983: Set connection var ansible_shell_type to sh 13731 1727203877.19989: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203877.19996: Set connection var ansible_connection to ssh 13731 1727203877.19999: Set connection var ansible_shell_executable to /bin/sh 13731 1727203877.20004: Set connection var ansible_timeout to 10 13731 1727203877.20017: variable 'ansible_shell_executable' from source: unknown 13731 1727203877.20020: variable 'ansible_connection' from source: unknown 13731 1727203877.20022: variable 'ansible_module_compression' from source: unknown 13731 1727203877.20024: variable 'ansible_shell_type' from source: unknown 13731 1727203877.20026: variable 'ansible_shell_executable' from source: unknown 13731 1727203877.20028: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203877.20033: variable 'ansible_pipelining' from source: unknown 13731 1727203877.20035: variable 'ansible_timeout' from source: unknown 13731 1727203877.20039: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203877.20103: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203877.20113: variable 'omit' from source: magic vars 13731 1727203877.20115: starting attempt loop 13731 1727203877.20118: running the handler 13731 1727203877.20121: _low_level_execute_command(): starting 13731 1727203877.20123: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203877.20549: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.20594: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203877.20597: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203877.20603: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.20606: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203877.20608: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.20610: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203877.20612: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.20650: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.20654: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.20662: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.20700: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.22263: stdout chunk (state=3): >>>/root <<< 13731 1727203877.22359: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.22398: stderr chunk (state=3): >>><<< 13731 1727203877.22401: stdout chunk (state=3): >>><<< 13731 1727203877.22409: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.22417: _low_level_execute_command(): starting 13731 1727203877.22423: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962 `" && echo ansible-tmp-1727203877.2240887-17865-63222882150962="` echo /root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962 `" ) && sleep 0' 13731 1727203877.22857: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.22869: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.22872: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203877.22874: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203877.22878: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.22965: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.23010: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.25083: stdout chunk (state=3): >>>ansible-tmp-1727203877.2240887-17865-63222882150962=/root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962 <<< 13731 1727203877.25218: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.25222: stdout chunk (state=3): >>><<< 13731 1727203877.25224: stderr chunk (state=3): >>><<< 13731 1727203877.25514: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203877.2240887-17865-63222882150962=/root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.25518: variable 'ansible_module_compression' from source: unknown 13731 1727203877.25521: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203877.25523: variable 'ansible_facts' from source: unknown 13731 1727203877.25614: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/AnsiballZ_command.py 13731 1727203877.25815: Sending initial data 13731 1727203877.25825: Sent initial data (155 bytes) 13731 1727203877.27225: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.27327: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.27643: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.27666: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.27735: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.29255: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203877.29389: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203877.29436: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpq38zr0da /root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/AnsiballZ_command.py <<< 13731 1727203877.29440: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/AnsiballZ_command.py" <<< 13731 1727203877.29558: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpq38zr0da" to remote "/root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/AnsiballZ_command.py" <<< 13731 1727203877.30890: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.30955: stderr chunk (state=3): >>><<< 13731 1727203877.30969: stdout chunk (state=3): >>><<< 13731 1727203877.31008: done transferring module to remote 13731 1727203877.31027: _low_level_execute_command(): starting 13731 1727203877.31036: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/ /root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/AnsiballZ_command.py && sleep 0' 13731 1727203877.31653: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203877.31670: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.31685: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.31702: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203877.31749: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.31815: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.31834: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.31851: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.31921: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.33882: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.33886: stdout chunk (state=3): >>><<< 13731 1727203877.33888: stderr chunk (state=3): >>><<< 13731 1727203877.33908: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.33916: _low_level_execute_command(): starting 13731 1727203877.33925: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/AnsiballZ_command.py && sleep 0' 13731 1727203877.34501: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203877.34515: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.34532: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.34549: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203877.34571: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203877.34588: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203877.34601: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.34616: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203877.34688: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.34707: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.34725: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.34739: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.34804: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.50124: stdout chunk (state=3): >>> {"changed": true, "stdout": "none 0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/arp_validate"], "start": "2024-09-24 14:51:17.497869", "end": "2024-09-24 14:51:17.500755", "delta": "0:00:00.002886", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/arp_validate", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203877.51544: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203877.51548: stdout chunk (state=3): >>><<< 13731 1727203877.51556: stderr chunk (state=3): >>><<< 13731 1727203877.51590: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "none 0", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/arp_validate"], "start": "2024-09-24 14:51:17.497869", "end": "2024-09-24 14:51:17.500755", "delta": "0:00:00.002886", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/arp_validate", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203877.51620: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/arp_validate', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203877.51625: _low_level_execute_command(): starting 13731 1727203877.51630: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203877.2240887-17865-63222882150962/ > /dev/null 2>&1 && sleep 0' 13731 1727203877.52781: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203877.52785: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.52788: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.52790: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203877.52792: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203877.52794: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203877.52796: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.52798: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203877.52800: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.52802: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.52985: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.52989: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.52991: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.54853: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.54856: stdout chunk (state=3): >>><<< 13731 1727203877.54869: stderr chunk (state=3): >>><<< 13731 1727203877.54883: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.54888: handler run complete 13731 1727203877.54903: Evaluated conditional (False): False 13731 1727203877.55011: variable 'bond_opt' from source: unknown 13731 1727203877.55016: variable 'result' from source: set_fact 13731 1727203877.55026: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203877.55035: attempt loop complete, returning result 13731 1727203877.55048: variable 'bond_opt' from source: unknown 13731 1727203877.55102: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'arp_validate', 'value': 'none'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "arp_validate", "value": "none" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/arp_validate" ], "delta": "0:00:00.002886", "end": "2024-09-24 14:51:17.500755", "rc": 0, "start": "2024-09-24 14:51:17.497869" } STDOUT: none 0 13731 1727203877.55229: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203877.55232: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203877.55234: variable 'omit' from source: magic vars 13731 1727203877.55323: variable 'ansible_distribution_major_version' from source: facts 13731 1727203877.55326: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203877.55330: variable 'omit' from source: magic vars 13731 1727203877.55341: variable 'omit' from source: magic vars 13731 1727203877.55449: variable 'controller_device' from source: play vars 13731 1727203877.55454: variable 'bond_opt' from source: unknown 13731 1727203877.55471: variable 'omit' from source: magic vars 13731 1727203877.55488: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203877.55495: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203877.55500: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203877.55512: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203877.55515: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203877.55518: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203877.55567: Set connection var ansible_pipelining to False 13731 1727203877.55574: Set connection var ansible_shell_type to sh 13731 1727203877.55586: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203877.55780: Set connection var ansible_connection to ssh 13731 1727203877.55783: Set connection var ansible_shell_executable to /bin/sh 13731 1727203877.55785: Set connection var ansible_timeout to 10 13731 1727203877.55787: variable 'ansible_shell_executable' from source: unknown 13731 1727203877.55789: variable 'ansible_connection' from source: unknown 13731 1727203877.55791: variable 'ansible_module_compression' from source: unknown 13731 1727203877.55793: variable 'ansible_shell_type' from source: unknown 13731 1727203877.55795: variable 'ansible_shell_executable' from source: unknown 13731 1727203877.55796: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203877.55798: variable 'ansible_pipelining' from source: unknown 13731 1727203877.55800: variable 'ansible_timeout' from source: unknown 13731 1727203877.55802: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203877.55804: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203877.55806: variable 'omit' from source: magic vars 13731 1727203877.55807: starting attempt loop 13731 1727203877.55809: running the handler 13731 1727203877.56211: _low_level_execute_command(): starting 13731 1727203877.56215: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203877.57307: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.57425: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.57466: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.57512: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.59142: stdout chunk (state=3): >>>/root <<< 13731 1727203877.59217: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.59221: stdout chunk (state=3): >>><<< 13731 1727203877.59223: stderr chunk (state=3): >>><<< 13731 1727203877.59310: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.59314: _low_level_execute_command(): starting 13731 1727203877.59316: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783 `" && echo ansible-tmp-1727203877.5924563-17865-77955011108783="` echo /root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783 `" ) && sleep 0' 13731 1727203877.60029: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203877.60092: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.60153: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.60169: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.60195: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.60352: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.62306: stdout chunk (state=3): >>>ansible-tmp-1727203877.5924563-17865-77955011108783=/root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783 <<< 13731 1727203877.62407: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.62410: stdout chunk (state=3): >>><<< 13731 1727203877.62412: stderr chunk (state=3): >>><<< 13731 1727203877.62414: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203877.5924563-17865-77955011108783=/root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.62641: variable 'ansible_module_compression' from source: unknown 13731 1727203877.62644: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203877.62646: variable 'ansible_facts' from source: unknown 13731 1727203877.62726: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/AnsiballZ_command.py 13731 1727203877.63202: Sending initial data 13731 1727203877.63211: Sent initial data (155 bytes) 13731 1727203877.64210: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.64224: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.64234: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.64287: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.64499: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.64583: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.66135: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203877.66265: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203877.66336: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp5auii208 /root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/AnsiballZ_command.py <<< 13731 1727203877.66344: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/AnsiballZ_command.py" <<< 13731 1727203877.66513: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp5auii208" to remote "/root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/AnsiballZ_command.py" <<< 13731 1727203877.67671: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.67674: stdout chunk (state=3): >>><<< 13731 1727203877.67684: stderr chunk (state=3): >>><<< 13731 1727203877.67727: done transferring module to remote 13731 1727203877.67735: _low_level_execute_command(): starting 13731 1727203877.67740: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/ /root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/AnsiballZ_command.py && sleep 0' 13731 1727203877.68771: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.68778: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203877.68793: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.68798: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203877.68811: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203877.68828: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.68833: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.68904: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.68908: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.68931: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.68989: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.70897: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.70901: stdout chunk (state=3): >>><<< 13731 1727203877.70903: stderr chunk (state=3): >>><<< 13731 1727203877.70906: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.70908: _low_level_execute_command(): starting 13731 1727203877.70910: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/AnsiballZ_command.py && sleep 0' 13731 1727203877.71703: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.71707: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.71714: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.71716: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.71795: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.71810: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.71825: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.71887: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.87547: stdout chunk (state=3): >>> {"changed": true, "stdout": "test1", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/primary"], "start": "2024-09-24 14:51:17.872176", "end": "2024-09-24 14:51:17.875122", "delta": "0:00:00.002946", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/primary", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203877.88977: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203877.89010: stderr chunk (state=3): >>><<< 13731 1727203877.89013: stdout chunk (state=3): >>><<< 13731 1727203877.89028: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "test1", "stderr": "", "rc": 0, "cmd": ["cat", "/sys/class/net/nm-bond/bonding/primary"], "start": "2024-09-24 14:51:17.872176", "end": "2024-09-24 14:51:17.875122", "delta": "0:00:00.002946", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /sys/class/net/nm-bond/bonding/primary", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203877.89049: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /sys/class/net/nm-bond/bonding/primary', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203877.89055: _low_level_execute_command(): starting 13731 1727203877.89060: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203877.5924563-17865-77955011108783/ > /dev/null 2>&1 && sleep 0' 13731 1727203877.89519: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203877.89523: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.89525: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203877.89529: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203877.89531: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203877.89581: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203877.89585: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203877.89596: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203877.89630: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203877.91412: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203877.91439: stderr chunk (state=3): >>><<< 13731 1727203877.91442: stdout chunk (state=3): >>><<< 13731 1727203877.91457: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203877.91462: handler run complete 13731 1727203877.91487: Evaluated conditional (False): False 13731 1727203877.91593: variable 'bond_opt' from source: unknown 13731 1727203877.91598: variable 'result' from source: set_fact 13731 1727203877.91610: Evaluated conditional (bond_opt.value in result.stdout): True 13731 1727203877.91619: attempt loop complete, returning result 13731 1727203877.91632: variable 'bond_opt' from source: unknown 13731 1727203877.91685: variable 'bond_opt' from source: unknown ok: [managed-node3] => (item={'key': 'primary', 'value': 'test1'}) => { "ansible_loop_var": "bond_opt", "attempts": 1, "bond_opt": { "key": "primary", "value": "test1" }, "changed": false, "cmd": [ "cat", "/sys/class/net/nm-bond/bonding/primary" ], "delta": "0:00:00.002946", "end": "2024-09-24 14:51:17.875122", "rc": 0, "start": "2024-09-24 14:51:17.872176" } STDOUT: test1 13731 1727203877.91813: dumping result to json 13731 1727203877.91816: done dumping result, returning 13731 1727203877.91818: done running TaskExecutor() for managed-node3/TASK: ** TEST check bond settings [028d2410-947f-82dc-c122-000000000c2a] 13731 1727203877.91820: sending task result for task 028d2410-947f-82dc-c122-000000000c2a 13731 1727203877.92014: no more pending results, returning what we have 13731 1727203877.92018: results queue empty 13731 1727203877.92019: checking for any_errors_fatal 13731 1727203877.92021: done checking for any_errors_fatal 13731 1727203877.92021: checking for max_fail_percentage 13731 1727203877.92025: done checking for max_fail_percentage 13731 1727203877.92026: checking to see if all hosts have failed and the running result is not ok 13731 1727203877.92027: done checking to see if all hosts have failed 13731 1727203877.92027: getting the remaining hosts for this loop 13731 1727203877.92029: done getting the remaining hosts for this loop 13731 1727203877.92032: getting the next task for host managed-node3 13731 1727203877.92039: done getting next task for host managed-node3 13731 1727203877.92042: ^ task is: TASK: Include the task 'assert_IPv4_present.yml' 13731 1727203877.92045: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203877.92048: getting variables 13731 1727203877.92049: in VariableManager get_vars() 13731 1727203877.92088: Calling all_inventory to load vars for managed-node3 13731 1727203877.92091: Calling groups_inventory to load vars for managed-node3 13731 1727203877.92099: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203877.92105: done sending task result for task 028d2410-947f-82dc-c122-000000000c2a 13731 1727203877.92108: WORKER PROCESS EXITING 13731 1727203877.92117: Calling all_plugins_play to load vars for managed-node3 13731 1727203877.92119: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203877.92122: Calling groups_plugins_play to load vars for managed-node3 13731 1727203877.93008: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203877.93878: done with get_vars() 13731 1727203877.93898: done getting variables TASK [Include the task 'assert_IPv4_present.yml'] ****************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml:11 Tuesday 24 September 2024 14:51:17 -0400 (0:00:01.908) 0:00:56.170 ***** 13731 1727203877.93971: entering _queue_task() for managed-node3/include_tasks 13731 1727203877.94235: worker is 1 (out of 1 available) 13731 1727203877.94249: exiting _queue_task() for managed-node3/include_tasks 13731 1727203877.94262: done queuing things up, now waiting for results queue to drain 13731 1727203877.94264: waiting for pending results... 13731 1727203877.94447: running TaskExecutor() for managed-node3/TASK: Include the task 'assert_IPv4_present.yml' 13731 1727203877.94533: in run() - task 028d2410-947f-82dc-c122-000000000c2c 13731 1727203877.94545: variable 'ansible_search_path' from source: unknown 13731 1727203877.94549: variable 'ansible_search_path' from source: unknown 13731 1727203877.94580: calling self._execute() 13731 1727203877.94659: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203877.94667: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203877.94677: variable 'omit' from source: magic vars 13731 1727203877.94953: variable 'ansible_distribution_major_version' from source: facts 13731 1727203877.94963: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203877.94971: _execute() done 13731 1727203877.94974: dumping result to json 13731 1727203877.94978: done dumping result, returning 13731 1727203877.94986: done running TaskExecutor() for managed-node3/TASK: Include the task 'assert_IPv4_present.yml' [028d2410-947f-82dc-c122-000000000c2c] 13731 1727203877.94990: sending task result for task 028d2410-947f-82dc-c122-000000000c2c 13731 1727203877.95077: done sending task result for task 028d2410-947f-82dc-c122-000000000c2c 13731 1727203877.95080: WORKER PROCESS EXITING 13731 1727203877.95109: no more pending results, returning what we have 13731 1727203877.95115: in VariableManager get_vars() 13731 1727203877.95163: Calling all_inventory to load vars for managed-node3 13731 1727203877.95166: Calling groups_inventory to load vars for managed-node3 13731 1727203877.95168: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203877.95182: Calling all_plugins_play to load vars for managed-node3 13731 1727203877.95184: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203877.95187: Calling groups_plugins_play to load vars for managed-node3 13731 1727203877.95988: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203877.96947: done with get_vars() 13731 1727203877.96960: variable 'ansible_search_path' from source: unknown 13731 1727203877.96962: variable 'ansible_search_path' from source: unknown 13731 1727203877.96968: variable 'item' from source: include params 13731 1727203877.97050: variable 'item' from source: include params 13731 1727203877.97078: we have included files to process 13731 1727203877.97079: generating all_blocks data 13731 1727203877.97080: done generating all_blocks data 13731 1727203877.97083: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml 13731 1727203877.97084: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml 13731 1727203877.97085: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml 13731 1727203877.97214: done processing included file 13731 1727203877.97215: iterating over new_blocks loaded from include file 13731 1727203877.97216: in VariableManager get_vars() 13731 1727203877.97232: done with get_vars() 13731 1727203877.97234: filtering new block on tags 13731 1727203877.97251: done filtering new block on tags 13731 1727203877.97252: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml for managed-node3 13731 1727203877.97256: extending task lists for all hosts with included blocks 13731 1727203877.97379: done extending task lists 13731 1727203877.97380: done processing included files 13731 1727203877.97381: results queue empty 13731 1727203877.97381: checking for any_errors_fatal 13731 1727203877.97388: done checking for any_errors_fatal 13731 1727203877.97388: checking for max_fail_percentage 13731 1727203877.97389: done checking for max_fail_percentage 13731 1727203877.97390: checking to see if all hosts have failed and the running result is not ok 13731 1727203877.97390: done checking to see if all hosts have failed 13731 1727203877.97390: getting the remaining hosts for this loop 13731 1727203877.97391: done getting the remaining hosts for this loop 13731 1727203877.97393: getting the next task for host managed-node3 13731 1727203877.97395: done getting next task for host managed-node3 13731 1727203877.97397: ^ task is: TASK: ** TEST check IPv4 13731 1727203877.97399: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203877.97400: getting variables 13731 1727203877.97401: in VariableManager get_vars() 13731 1727203877.97409: Calling all_inventory to load vars for managed-node3 13731 1727203877.97411: Calling groups_inventory to load vars for managed-node3 13731 1727203877.97412: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203877.97416: Calling all_plugins_play to load vars for managed-node3 13731 1727203877.97417: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203877.97419: Calling groups_plugins_play to load vars for managed-node3 13731 1727203877.98053: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203877.98887: done with get_vars() 13731 1727203877.98901: done getting variables 13731 1727203877.98929: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [** TEST check IPv4] ****************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml:3 Tuesday 24 September 2024 14:51:17 -0400 (0:00:00.049) 0:00:56.220 ***** 13731 1727203877.98950: entering _queue_task() for managed-node3/command 13731 1727203877.99199: worker is 1 (out of 1 available) 13731 1727203877.99211: exiting _queue_task() for managed-node3/command 13731 1727203877.99223: done queuing things up, now waiting for results queue to drain 13731 1727203877.99225: waiting for pending results... 13731 1727203877.99411: running TaskExecutor() for managed-node3/TASK: ** TEST check IPv4 13731 1727203877.99501: in run() - task 028d2410-947f-82dc-c122-000000000da6 13731 1727203877.99514: variable 'ansible_search_path' from source: unknown 13731 1727203877.99518: variable 'ansible_search_path' from source: unknown 13731 1727203877.99548: calling self._execute() 13731 1727203877.99626: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203877.99630: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203877.99639: variable 'omit' from source: magic vars 13731 1727203877.99909: variable 'ansible_distribution_major_version' from source: facts 13731 1727203877.99918: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203877.99923: variable 'omit' from source: magic vars 13731 1727203877.99955: variable 'omit' from source: magic vars 13731 1727203878.00113: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203878.01522: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203878.01568: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203878.01599: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203878.01624: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203878.01646: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203878.01705: variable 'interface' from source: include params 13731 1727203878.01709: variable 'controller_device' from source: play vars 13731 1727203878.01756: variable 'controller_device' from source: play vars 13731 1727203878.01774: variable 'omit' from source: magic vars 13731 1727203878.01798: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203878.01818: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203878.01832: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203878.01845: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203878.01863: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203878.01882: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203878.01885: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203878.01888: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203878.01950: Set connection var ansible_pipelining to False 13731 1727203878.01953: Set connection var ansible_shell_type to sh 13731 1727203878.01959: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203878.01969: Set connection var ansible_connection to ssh 13731 1727203878.01973: Set connection var ansible_shell_executable to /bin/sh 13731 1727203878.01977: Set connection var ansible_timeout to 10 13731 1727203878.01994: variable 'ansible_shell_executable' from source: unknown 13731 1727203878.01997: variable 'ansible_connection' from source: unknown 13731 1727203878.02000: variable 'ansible_module_compression' from source: unknown 13731 1727203878.02002: variable 'ansible_shell_type' from source: unknown 13731 1727203878.02004: variable 'ansible_shell_executable' from source: unknown 13731 1727203878.02006: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203878.02008: variable 'ansible_pipelining' from source: unknown 13731 1727203878.02010: variable 'ansible_timeout' from source: unknown 13731 1727203878.02015: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203878.02181: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203878.02185: variable 'omit' from source: magic vars 13731 1727203878.02187: starting attempt loop 13731 1727203878.02189: running the handler 13731 1727203878.02192: _low_level_execute_command(): starting 13731 1727203878.02194: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203878.02792: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203878.02808: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203878.02828: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203878.02901: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203878.04662: stdout chunk (state=3): >>>/root <<< 13731 1727203878.04760: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203878.04790: stderr chunk (state=3): >>><<< 13731 1727203878.04797: stdout chunk (state=3): >>><<< 13731 1727203878.04816: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203878.04828: _low_level_execute_command(): starting 13731 1727203878.04836: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530 `" && echo ansible-tmp-1727203878.0482087-17972-144642051830530="` echo /root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530 `" ) && sleep 0' 13731 1727203878.05278: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203878.05346: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203878.05351: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203878.05415: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203878.05431: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203878.07270: stdout chunk (state=3): >>>ansible-tmp-1727203878.0482087-17972-144642051830530=/root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530 <<< 13731 1727203878.07415: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203878.07431: stderr chunk (state=3): >>><<< 13731 1727203878.07445: stdout chunk (state=3): >>><<< 13731 1727203878.07482: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203878.0482087-17972-144642051830530=/root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203878.07505: variable 'ansible_module_compression' from source: unknown 13731 1727203878.07566: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203878.07638: variable 'ansible_facts' from source: unknown 13731 1727203878.07705: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/AnsiballZ_command.py 13731 1727203878.07937: Sending initial data 13731 1727203878.07949: Sent initial data (156 bytes) 13731 1727203878.08500: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203878.08603: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203878.08627: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203878.08644: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203878.08675: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203878.08743: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203878.10243: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203878.10292: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203878.10332: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpboucsvbg /root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/AnsiballZ_command.py <<< 13731 1727203878.10350: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/AnsiballZ_command.py" <<< 13731 1727203878.10382: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpboucsvbg" to remote "/root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/AnsiballZ_command.py" <<< 13731 1727203878.11083: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203878.11115: stderr chunk (state=3): >>><<< 13731 1727203878.11182: stdout chunk (state=3): >>><<< 13731 1727203878.11189: done transferring module to remote 13731 1727203878.11203: _low_level_execute_command(): starting 13731 1727203878.11211: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/ /root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/AnsiballZ_command.py && sleep 0' 13731 1727203878.11818: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203878.11834: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203878.11853: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203878.11872: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203878.11890: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203878.11952: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203878.12003: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203878.12018: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203878.12058: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203878.12116: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203878.14085: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203878.14088: stdout chunk (state=3): >>><<< 13731 1727203878.14091: stderr chunk (state=3): >>><<< 13731 1727203878.14107: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203878.14185: _low_level_execute_command(): starting 13731 1727203878.14192: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/AnsiballZ_command.py && sleep 0' 13731 1727203878.14779: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203878.14854: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203878.14908: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203878.14926: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203878.14946: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203878.15014: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203878.31070: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "", "rc": 0, "cmd": ["ip", "-4", "a", "s", "nm-bond"], "start": "2024-09-24 14:51:18.303279", "end": "2024-09-24 14:51:18.306892", "delta": "0:00:00.003613", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -4 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203878.33105: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203878.33494: stderr chunk (state=3): >>><<< 13731 1727203878.33503: stdout chunk (state=3): >>><<< 13731 1727203878.33869: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "", "rc": 0, "cmd": ["ip", "-4", "a", "s", "nm-bond"], "start": "2024-09-24 14:51:18.303279", "end": "2024-09-24 14:51:18.306892", "delta": "0:00:00.003613", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -4 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203878.33872: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip -4 a s nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203878.33878: _low_level_execute_command(): starting 13731 1727203878.33880: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203878.0482087-17972-144642051830530/ > /dev/null 2>&1 && sleep 0' 13731 1727203878.35512: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203878.35787: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203878.36002: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203878.36017: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203878.36207: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203878.38093: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203878.38097: stdout chunk (state=3): >>><<< 13731 1727203878.38099: stderr chunk (state=3): >>><<< 13731 1727203878.38101: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203878.38103: handler run complete 13731 1727203878.38105: Evaluated conditional (False): False 13731 1727203878.38184: variable 'address' from source: include params 13731 1727203878.38199: variable 'result' from source: set_fact 13731 1727203878.38220: Evaluated conditional (address in result.stdout): False 13731 1727203878.38226: Retrying task, attempt 1 of 21 FAILED - RETRYING: [managed-node3]: ** TEST check IPv4 (20 retries left). 13731 1727203880.38508: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203880.38513: running the handler 13731 1727203880.38520: _low_level_execute_command(): starting 13731 1727203880.38525: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203880.38990: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203880.38995: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.38998: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203880.39001: stderr chunk (state=3): >>>debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.39054: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203880.39058: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.39105: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.40766: stdout chunk (state=3): >>>/root <<< 13731 1727203880.40861: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.40894: stderr chunk (state=3): >>><<< 13731 1727203880.40898: stdout chunk (state=3): >>><<< 13731 1727203880.40911: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203880.40919: _low_level_execute_command(): starting 13731 1727203880.40925: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675 `" && echo ansible-tmp-1727203880.4091098-17972-204616154294675="` echo /root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675 `" ) && sleep 0' 13731 1727203880.41352: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203880.41355: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203880.41357: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.41360: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203880.41364: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.41419: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203880.41424: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203880.41434: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.41457: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.43580: stdout chunk (state=3): >>>ansible-tmp-1727203880.4091098-17972-204616154294675=/root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675 <<< 13731 1727203880.43740: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.43748: stdout chunk (state=3): >>><<< 13731 1727203880.43751: stderr chunk (state=3): >>><<< 13731 1727203880.43770: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203880.4091098-17972-204616154294675=/root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203880.43803: variable 'ansible_module_compression' from source: unknown 13731 1727203880.43951: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203880.43955: variable 'ansible_facts' from source: unknown 13731 1727203880.43971: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/AnsiballZ_command.py 13731 1727203880.44095: Sending initial data 13731 1727203880.44191: Sent initial data (156 bytes) 13731 1727203880.44798: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203880.44815: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203880.44869: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.44944: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203880.44988: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203880.45005: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.45084: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.46580: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203880.46623: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203880.46660: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpseaagdq6 /root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/AnsiballZ_command.py <<< 13731 1727203880.46663: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/AnsiballZ_command.py" <<< 13731 1727203880.46707: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpseaagdq6" to remote "/root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/AnsiballZ_command.py" <<< 13731 1727203880.47432: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.47532: stdout chunk (state=3): >>><<< 13731 1727203880.47535: stderr chunk (state=3): >>><<< 13731 1727203880.47537: done transferring module to remote 13731 1727203880.47539: _low_level_execute_command(): starting 13731 1727203880.47565: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/ /root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/AnsiballZ_command.py && sleep 0' 13731 1727203880.48024: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203880.48037: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203880.48047: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.48096: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203880.48108: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.48146: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.50074: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.50080: stdout chunk (state=3): >>><<< 13731 1727203880.50082: stderr chunk (state=3): >>><<< 13731 1727203880.50097: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203880.50174: _low_level_execute_command(): starting 13731 1727203880.50179: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/AnsiballZ_command.py && sleep 0' 13731 1727203880.50707: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203880.50726: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203880.50740: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203880.50840: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203880.50856: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203880.50872: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.50958: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.66348: stdout chunk (state=3): >>> {"changed": true, "stdout": "24: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet 192.0.2.95/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond\n valid_lft 237sec preferred_lft 237sec", "stderr": "", "rc": 0, "cmd": ["ip", "-4", "a", "s", "nm-bond"], "start": "2024-09-24 14:51:20.659458", "end": "2024-09-24 14:51:20.662989", "delta": "0:00:00.003531", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -4 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203880.68064: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203880.68068: stdout chunk (state=3): >>><<< 13731 1727203880.68070: stderr chunk (state=3): >>><<< 13731 1727203880.68087: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "24: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet 192.0.2.95/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond\n valid_lft 237sec preferred_lft 237sec", "stderr": "", "rc": 0, "cmd": ["ip", "-4", "a", "s", "nm-bond"], "start": "2024-09-24 14:51:20.659458", "end": "2024-09-24 14:51:20.662989", "delta": "0:00:00.003531", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -4 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203880.68121: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip -4 a s nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203880.68126: _low_level_execute_command(): starting 13731 1727203880.68131: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203880.4091098-17972-204616154294675/ > /dev/null 2>&1 && sleep 0' 13731 1727203880.68981: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203880.68989: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203880.68992: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203880.68994: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203880.68997: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203880.68999: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203880.69001: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.69002: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203880.69005: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203880.69007: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203880.69009: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203880.69011: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.69014: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.69059: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.70809: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.70839: stderr chunk (state=3): >>><<< 13731 1727203880.70842: stdout chunk (state=3): >>><<< 13731 1727203880.70858: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203880.70867: handler run complete 13731 1727203880.70887: Evaluated conditional (False): False 13731 1727203880.71021: variable 'address' from source: include params 13731 1727203880.71024: variable 'result' from source: set_fact 13731 1727203880.71181: Evaluated conditional (address in result.stdout): True 13731 1727203880.71184: attempt loop complete, returning result 13731 1727203880.71186: _execute() done 13731 1727203880.71187: dumping result to json 13731 1727203880.71189: done dumping result, returning 13731 1727203880.71190: done running TaskExecutor() for managed-node3/TASK: ** TEST check IPv4 [028d2410-947f-82dc-c122-000000000da6] 13731 1727203880.71192: sending task result for task 028d2410-947f-82dc-c122-000000000da6 13731 1727203880.71239: done sending task result for task 028d2410-947f-82dc-c122-000000000da6 13731 1727203880.71242: WORKER PROCESS EXITING ok: [managed-node3] => { "attempts": 2, "changed": false, "cmd": [ "ip", "-4", "a", "s", "nm-bond" ], "delta": "0:00:00.003531", "end": "2024-09-24 14:51:20.662989", "rc": 0, "start": "2024-09-24 14:51:20.659458" } STDOUT: 24: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000 inet 192.0.2.95/24 brd 192.0.2.255 scope global dynamic noprefixroute nm-bond valid_lft 237sec preferred_lft 237sec 13731 1727203880.71309: no more pending results, returning what we have 13731 1727203880.71313: results queue empty 13731 1727203880.71314: checking for any_errors_fatal 13731 1727203880.71315: done checking for any_errors_fatal 13731 1727203880.71316: checking for max_fail_percentage 13731 1727203880.71317: done checking for max_fail_percentage 13731 1727203880.71318: checking to see if all hosts have failed and the running result is not ok 13731 1727203880.71319: done checking to see if all hosts have failed 13731 1727203880.71319: getting the remaining hosts for this loop 13731 1727203880.71321: done getting the remaining hosts for this loop 13731 1727203880.71324: getting the next task for host managed-node3 13731 1727203880.71333: done getting next task for host managed-node3 13731 1727203880.71336: ^ task is: TASK: Include the task 'assert_IPv6_present.yml' 13731 1727203880.71339: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203880.71342: getting variables 13731 1727203880.71344: in VariableManager get_vars() 13731 1727203880.71384: Calling all_inventory to load vars for managed-node3 13731 1727203880.71387: Calling groups_inventory to load vars for managed-node3 13731 1727203880.71389: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203880.71398: Calling all_plugins_play to load vars for managed-node3 13731 1727203880.71400: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203880.71403: Calling groups_plugins_play to load vars for managed-node3 13731 1727203880.72965: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203880.74703: done with get_vars() 13731 1727203880.74729: done getting variables TASK [Include the task 'assert_IPv6_present.yml'] ****************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml:16 Tuesday 24 September 2024 14:51:20 -0400 (0:00:02.758) 0:00:58.979 ***** 13731 1727203880.74830: entering _queue_task() for managed-node3/include_tasks 13731 1727203880.75166: worker is 1 (out of 1 available) 13731 1727203880.75384: exiting _queue_task() for managed-node3/include_tasks 13731 1727203880.75394: done queuing things up, now waiting for results queue to drain 13731 1727203880.75396: waiting for pending results... 13731 1727203880.75468: running TaskExecutor() for managed-node3/TASK: Include the task 'assert_IPv6_present.yml' 13731 1727203880.75620: in run() - task 028d2410-947f-82dc-c122-000000000c2d 13731 1727203880.75625: variable 'ansible_search_path' from source: unknown 13731 1727203880.75629: variable 'ansible_search_path' from source: unknown 13731 1727203880.75658: calling self._execute() 13731 1727203880.75764: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203880.75836: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203880.75840: variable 'omit' from source: magic vars 13731 1727203880.76157: variable 'ansible_distribution_major_version' from source: facts 13731 1727203880.76180: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203880.76191: _execute() done 13731 1727203880.76199: dumping result to json 13731 1727203880.76205: done dumping result, returning 13731 1727203880.76214: done running TaskExecutor() for managed-node3/TASK: Include the task 'assert_IPv6_present.yml' [028d2410-947f-82dc-c122-000000000c2d] 13731 1727203880.76223: sending task result for task 028d2410-947f-82dc-c122-000000000c2d 13731 1727203880.76462: no more pending results, returning what we have 13731 1727203880.76469: in VariableManager get_vars() 13731 1727203880.76522: Calling all_inventory to load vars for managed-node3 13731 1727203880.76525: Calling groups_inventory to load vars for managed-node3 13731 1727203880.76528: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203880.76542: Calling all_plugins_play to load vars for managed-node3 13731 1727203880.76546: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203880.76549: Calling groups_plugins_play to load vars for managed-node3 13731 1727203880.77189: done sending task result for task 028d2410-947f-82dc-c122-000000000c2d 13731 1727203880.77193: WORKER PROCESS EXITING 13731 1727203880.78055: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203880.79547: done with get_vars() 13731 1727203880.79571: variable 'ansible_search_path' from source: unknown 13731 1727203880.79572: variable 'ansible_search_path' from source: unknown 13731 1727203880.79583: variable 'item' from source: include params 13731 1727203880.79687: variable 'item' from source: include params 13731 1727203880.79721: we have included files to process 13731 1727203880.79723: generating all_blocks data 13731 1727203880.79725: done generating all_blocks data 13731 1727203880.79729: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml 13731 1727203880.79731: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml 13731 1727203880.79733: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml 13731 1727203880.79922: done processing included file 13731 1727203880.79924: iterating over new_blocks loaded from include file 13731 1727203880.79925: in VariableManager get_vars() 13731 1727203880.79941: done with get_vars() 13731 1727203880.79942: filtering new block on tags 13731 1727203880.79959: done filtering new block on tags 13731 1727203880.79960: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml for managed-node3 13731 1727203880.79967: extending task lists for all hosts with included blocks 13731 1727203880.80190: done extending task lists 13731 1727203880.80192: done processing included files 13731 1727203880.80192: results queue empty 13731 1727203880.80193: checking for any_errors_fatal 13731 1727203880.80197: done checking for any_errors_fatal 13731 1727203880.80197: checking for max_fail_percentage 13731 1727203880.80198: done checking for max_fail_percentage 13731 1727203880.80198: checking to see if all hosts have failed and the running result is not ok 13731 1727203880.80199: done checking to see if all hosts have failed 13731 1727203880.80199: getting the remaining hosts for this loop 13731 1727203880.80200: done getting the remaining hosts for this loop 13731 1727203880.80202: getting the next task for host managed-node3 13731 1727203880.80204: done getting next task for host managed-node3 13731 1727203880.80206: ^ task is: TASK: ** TEST check IPv6 13731 1727203880.80208: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203880.80210: getting variables 13731 1727203880.80210: in VariableManager get_vars() 13731 1727203880.80220: Calling all_inventory to load vars for managed-node3 13731 1727203880.80221: Calling groups_inventory to load vars for managed-node3 13731 1727203880.80222: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203880.80226: Calling all_plugins_play to load vars for managed-node3 13731 1727203880.80227: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203880.80229: Calling groups_plugins_play to load vars for managed-node3 13731 1727203880.80933: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203880.81821: done with get_vars() 13731 1727203880.81841: done getting variables 13731 1727203880.81886: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [** TEST check IPv6] ****************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv6_present.yml:3 Tuesday 24 September 2024 14:51:20 -0400 (0:00:00.070) 0:00:59.050 ***** 13731 1727203880.81917: entering _queue_task() for managed-node3/command 13731 1727203880.82504: worker is 1 (out of 1 available) 13731 1727203880.82512: exiting _queue_task() for managed-node3/command 13731 1727203880.82523: done queuing things up, now waiting for results queue to drain 13731 1727203880.82525: waiting for pending results... 13731 1727203880.82581: running TaskExecutor() for managed-node3/TASK: ** TEST check IPv6 13731 1727203880.82724: in run() - task 028d2410-947f-82dc-c122-000000000dc9 13731 1727203880.82749: variable 'ansible_search_path' from source: unknown 13731 1727203880.82760: variable 'ansible_search_path' from source: unknown 13731 1727203880.82807: calling self._execute() 13731 1727203880.82916: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203880.82919: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203880.82928: variable 'omit' from source: magic vars 13731 1727203880.83210: variable 'ansible_distribution_major_version' from source: facts 13731 1727203880.83220: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203880.83225: variable 'omit' from source: magic vars 13731 1727203880.83263: variable 'omit' from source: magic vars 13731 1727203880.83380: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203880.84868: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203880.85181: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203880.85185: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203880.85187: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203880.85190: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203880.85192: variable 'controller_device' from source: play vars 13731 1727203880.85194: variable 'omit' from source: magic vars 13731 1727203880.85197: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203880.85216: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203880.85236: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203880.85257: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203880.85278: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203880.85314: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203880.85323: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203880.85332: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203880.85436: Set connection var ansible_pipelining to False 13731 1727203880.85447: Set connection var ansible_shell_type to sh 13731 1727203880.85459: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203880.85474: Set connection var ansible_connection to ssh 13731 1727203880.85488: Set connection var ansible_shell_executable to /bin/sh 13731 1727203880.85497: Set connection var ansible_timeout to 10 13731 1727203880.85531: variable 'ansible_shell_executable' from source: unknown 13731 1727203880.85539: variable 'ansible_connection' from source: unknown 13731 1727203880.85547: variable 'ansible_module_compression' from source: unknown 13731 1727203880.85557: variable 'ansible_shell_type' from source: unknown 13731 1727203880.85565: variable 'ansible_shell_executable' from source: unknown 13731 1727203880.85572: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203880.85583: variable 'ansible_pipelining' from source: unknown 13731 1727203880.85589: variable 'ansible_timeout' from source: unknown 13731 1727203880.85596: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203880.85718: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203880.85743: variable 'omit' from source: magic vars 13731 1727203880.85755: starting attempt loop 13731 1727203880.85761: running the handler 13731 1727203880.85784: _low_level_execute_command(): starting 13731 1727203880.85794: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203880.86459: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203880.86481: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.86541: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203880.86553: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.86597: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.88225: stdout chunk (state=3): >>>/root <<< 13731 1727203880.88355: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.88385: stderr chunk (state=3): >>><<< 13731 1727203880.88391: stdout chunk (state=3): >>><<< 13731 1727203880.88410: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203880.88505: _low_level_execute_command(): starting 13731 1727203880.88512: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837 `" && echo ansible-tmp-1727203880.8841584-18092-227120248194837="` echo /root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837 `" ) && sleep 0' 13731 1727203880.89040: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.89117: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.89174: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.91017: stdout chunk (state=3): >>>ansible-tmp-1727203880.8841584-18092-227120248194837=/root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837 <<< 13731 1727203880.91129: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.91152: stderr chunk (state=3): >>><<< 13731 1727203880.91155: stdout chunk (state=3): >>><<< 13731 1727203880.91172: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203880.8841584-18092-227120248194837=/root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203880.91197: variable 'ansible_module_compression' from source: unknown 13731 1727203880.91240: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203880.91266: variable 'ansible_facts' from source: unknown 13731 1727203880.91325: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/AnsiballZ_command.py 13731 1727203880.91431: Sending initial data 13731 1727203880.91435: Sent initial data (156 bytes) 13731 1727203880.92095: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.92151: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203880.92168: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203880.92190: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.92255: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.93759: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203880.93788: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203880.93822: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpr9w4q491 /root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/AnsiballZ_command.py <<< 13731 1727203880.93835: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/AnsiballZ_command.py" <<< 13731 1727203880.93863: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpr9w4q491" to remote "/root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/AnsiballZ_command.py" <<< 13731 1727203880.93865: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/AnsiballZ_command.py" <<< 13731 1727203880.94374: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.94417: stderr chunk (state=3): >>><<< 13731 1727203880.94420: stdout chunk (state=3): >>><<< 13731 1727203880.94435: done transferring module to remote 13731 1727203880.94445: _low_level_execute_command(): starting 13731 1727203880.94449: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/ /root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/AnsiballZ_command.py && sleep 0' 13731 1727203880.94886: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203880.94889: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203880.94892: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.94894: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203880.94899: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203880.94901: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.94950: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203880.94955: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.94987: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203880.96685: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203880.96709: stderr chunk (state=3): >>><<< 13731 1727203880.96712: stdout chunk (state=3): >>><<< 13731 1727203880.96725: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203880.96728: _low_level_execute_command(): starting 13731 1727203880.96733: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/AnsiballZ_command.py && sleep 0' 13731 1727203880.97138: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203880.97142: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.97154: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203880.97210: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203880.97217: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203880.97251: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203881.12814: stdout chunk (state=3): >>> {"changed": true, "stdout": "24: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet6 2001:db8::b9/128 scope global dynamic noprefixroute \n valid_lft 237sec preferred_lft 237sec\n inet6 2001:db8::407:85ff:fe66:e341/64 scope global dynamic noprefixroute \n valid_lft 1800sec preferred_lft 1800sec\n inet6 fe80::407:85ff:fe66:e341/64 scope link noprefixroute \n valid_lft forever preferred_lft forever", "stderr": "", "rc": 0, "cmd": ["ip", "-6", "a", "s", "nm-bond"], "start": "2024-09-24 14:51:21.124138", "end": "2024-09-24 14:51:21.127727", "delta": "0:00:00.003589", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -6 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203881.14292: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203881.14322: stderr chunk (state=3): >>><<< 13731 1727203881.14325: stdout chunk (state=3): >>><<< 13731 1727203881.14340: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "24: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000\n inet6 2001:db8::b9/128 scope global dynamic noprefixroute \n valid_lft 237sec preferred_lft 237sec\n inet6 2001:db8::407:85ff:fe66:e341/64 scope global dynamic noprefixroute \n valid_lft 1800sec preferred_lft 1800sec\n inet6 fe80::407:85ff:fe66:e341/64 scope link noprefixroute \n valid_lft forever preferred_lft forever", "stderr": "", "rc": 0, "cmd": ["ip", "-6", "a", "s", "nm-bond"], "start": "2024-09-24 14:51:21.124138", "end": "2024-09-24 14:51:21.127727", "delta": "0:00:00.003589", "msg": "", "invocation": {"module_args": {"_raw_params": "ip -6 a s nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203881.14374: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip -6 a s nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203881.14385: _low_level_execute_command(): starting 13731 1727203881.14388: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203880.8841584-18092-227120248194837/ > /dev/null 2>&1 && sleep 0' 13731 1727203881.14844: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203881.14847: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203881.14849: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.14852: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203881.14854: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.14908: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203881.14911: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203881.14948: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203881.16992: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203881.17015: stderr chunk (state=3): >>><<< 13731 1727203881.17018: stdout chunk (state=3): >>><<< 13731 1727203881.17033: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203881.17042: handler run complete 13731 1727203881.17066: Evaluated conditional (False): False 13731 1727203881.17172: variable 'address' from source: include params 13731 1727203881.17177: variable 'result' from source: set_fact 13731 1727203881.17191: Evaluated conditional (address in result.stdout): True 13731 1727203881.17201: attempt loop complete, returning result 13731 1727203881.17203: _execute() done 13731 1727203881.17206: dumping result to json 13731 1727203881.17212: done dumping result, returning 13731 1727203881.17219: done running TaskExecutor() for managed-node3/TASK: ** TEST check IPv6 [028d2410-947f-82dc-c122-000000000dc9] 13731 1727203881.17223: sending task result for task 028d2410-947f-82dc-c122-000000000dc9 13731 1727203881.17322: done sending task result for task 028d2410-947f-82dc-c122-000000000dc9 13731 1727203881.17325: WORKER PROCESS EXITING ok: [managed-node3] => { "attempts": 1, "changed": false, "cmd": [ "ip", "-6", "a", "s", "nm-bond" ], "delta": "0:00:00.003589", "end": "2024-09-24 14:51:21.127727", "rc": 0, "start": "2024-09-24 14:51:21.124138" } STDOUT: 24: nm-bond: mtu 1500 qdisc noqueue state UP group default qlen 1000 inet6 2001:db8::b9/128 scope global dynamic noprefixroute valid_lft 237sec preferred_lft 237sec inet6 2001:db8::407:85ff:fe66:e341/64 scope global dynamic noprefixroute valid_lft 1800sec preferred_lft 1800sec inet6 fe80::407:85ff:fe66:e341/64 scope link noprefixroute valid_lft forever preferred_lft forever 13731 1727203881.17405: no more pending results, returning what we have 13731 1727203881.17409: results queue empty 13731 1727203881.17410: checking for any_errors_fatal 13731 1727203881.17411: done checking for any_errors_fatal 13731 1727203881.17412: checking for max_fail_percentage 13731 1727203881.17414: done checking for max_fail_percentage 13731 1727203881.17415: checking to see if all hosts have failed and the running result is not ok 13731 1727203881.17415: done checking to see if all hosts have failed 13731 1727203881.17416: getting the remaining hosts for this loop 13731 1727203881.17418: done getting the remaining hosts for this loop 13731 1727203881.17421: getting the next task for host managed-node3 13731 1727203881.17431: done getting next task for host managed-node3 13731 1727203881.17441: ^ task is: TASK: Conditional asserts 13731 1727203881.17443: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203881.17448: getting variables 13731 1727203881.17450: in VariableManager get_vars() 13731 1727203881.17495: Calling all_inventory to load vars for managed-node3 13731 1727203881.17498: Calling groups_inventory to load vars for managed-node3 13731 1727203881.17500: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.17510: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.17513: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.17515: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.18321: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.19171: done with get_vars() 13731 1727203881.19191: done getting variables TASK [Conditional asserts] ***************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:42 Tuesday 24 September 2024 14:51:21 -0400 (0:00:00.373) 0:00:59.423 ***** 13731 1727203881.19258: entering _queue_task() for managed-node3/include_tasks 13731 1727203881.19508: worker is 1 (out of 1 available) 13731 1727203881.19523: exiting _queue_task() for managed-node3/include_tasks 13731 1727203881.19535: done queuing things up, now waiting for results queue to drain 13731 1727203881.19536: waiting for pending results... 13731 1727203881.19720: running TaskExecutor() for managed-node3/TASK: Conditional asserts 13731 1727203881.19797: in run() - task 028d2410-947f-82dc-c122-0000000008f0 13731 1727203881.19814: variable 'ansible_search_path' from source: unknown 13731 1727203881.19818: variable 'ansible_search_path' from source: unknown 13731 1727203881.20028: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203881.21734: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203881.21789: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203881.21815: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203881.21842: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203881.21863: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203881.21923: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203881.21949: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203881.21968: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203881.21996: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203881.22006: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203881.22118: dumping result to json 13731 1727203881.22121: done dumping result, returning 13731 1727203881.22124: done running TaskExecutor() for managed-node3/TASK: Conditional asserts [028d2410-947f-82dc-c122-0000000008f0] 13731 1727203881.22130: sending task result for task 028d2410-947f-82dc-c122-0000000008f0 13731 1727203881.22217: done sending task result for task 028d2410-947f-82dc-c122-0000000008f0 13731 1727203881.22220: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "skipped_reason": "No items in the list" } 13731 1727203881.22267: no more pending results, returning what we have 13731 1727203881.22272: results queue empty 13731 1727203881.22272: checking for any_errors_fatal 13731 1727203881.22284: done checking for any_errors_fatal 13731 1727203881.22285: checking for max_fail_percentage 13731 1727203881.22286: done checking for max_fail_percentage 13731 1727203881.22287: checking to see if all hosts have failed and the running result is not ok 13731 1727203881.22288: done checking to see if all hosts have failed 13731 1727203881.22288: getting the remaining hosts for this loop 13731 1727203881.22290: done getting the remaining hosts for this loop 13731 1727203881.22294: getting the next task for host managed-node3 13731 1727203881.22301: done getting next task for host managed-node3 13731 1727203881.22304: ^ task is: TASK: Success in test '{{ lsr_description }}' 13731 1727203881.22306: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203881.22310: getting variables 13731 1727203881.22312: in VariableManager get_vars() 13731 1727203881.22355: Calling all_inventory to load vars for managed-node3 13731 1727203881.22357: Calling groups_inventory to load vars for managed-node3 13731 1727203881.22360: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.22369: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.22372: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.22375: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.23263: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.24104: done with get_vars() 13731 1727203881.24123: done getting variables 13731 1727203881.24164: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203881.24249: variable 'lsr_description' from source: include params TASK [Success in test 'Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.'] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:47 Tuesday 24 September 2024 14:51:21 -0400 (0:00:00.050) 0:00:59.473 ***** 13731 1727203881.24272: entering _queue_task() for managed-node3/debug 13731 1727203881.24507: worker is 1 (out of 1 available) 13731 1727203881.24521: exiting _queue_task() for managed-node3/debug 13731 1727203881.24533: done queuing things up, now waiting for results queue to drain 13731 1727203881.24535: waiting for pending results... 13731 1727203881.24722: running TaskExecutor() for managed-node3/TASK: Success in test 'Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.' 13731 1727203881.24793: in run() - task 028d2410-947f-82dc-c122-0000000008f1 13731 1727203881.24808: variable 'ansible_search_path' from source: unknown 13731 1727203881.24811: variable 'ansible_search_path' from source: unknown 13731 1727203881.24841: calling self._execute() 13731 1727203881.24920: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.24924: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.24932: variable 'omit' from source: magic vars 13731 1727203881.25196: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.25207: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.25211: variable 'omit' from source: magic vars 13731 1727203881.25238: variable 'omit' from source: magic vars 13731 1727203881.25308: variable 'lsr_description' from source: include params 13731 1727203881.25322: variable 'omit' from source: magic vars 13731 1727203881.25354: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203881.25382: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203881.25399: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203881.25412: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203881.25428: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203881.25446: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203881.25449: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.25452: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.25522: Set connection var ansible_pipelining to False 13731 1727203881.25528: Set connection var ansible_shell_type to sh 13731 1727203881.25530: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203881.25540: Set connection var ansible_connection to ssh 13731 1727203881.25544: Set connection var ansible_shell_executable to /bin/sh 13731 1727203881.25547: Set connection var ansible_timeout to 10 13731 1727203881.25562: variable 'ansible_shell_executable' from source: unknown 13731 1727203881.25567: variable 'ansible_connection' from source: unknown 13731 1727203881.25569: variable 'ansible_module_compression' from source: unknown 13731 1727203881.25572: variable 'ansible_shell_type' from source: unknown 13731 1727203881.25574: variable 'ansible_shell_executable' from source: unknown 13731 1727203881.25578: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.25583: variable 'ansible_pipelining' from source: unknown 13731 1727203881.25585: variable 'ansible_timeout' from source: unknown 13731 1727203881.25589: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.25691: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203881.25700: variable 'omit' from source: magic vars 13731 1727203881.25705: starting attempt loop 13731 1727203881.25708: running the handler 13731 1727203881.25743: handler run complete 13731 1727203881.25758: attempt loop complete, returning result 13731 1727203881.25761: _execute() done 13731 1727203881.25764: dumping result to json 13731 1727203881.25767: done dumping result, returning 13731 1727203881.25773: done running TaskExecutor() for managed-node3/TASK: Success in test 'Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.' [028d2410-947f-82dc-c122-0000000008f1] 13731 1727203881.25778: sending task result for task 028d2410-947f-82dc-c122-0000000008f1 13731 1727203881.25850: done sending task result for task 028d2410-947f-82dc-c122-0000000008f1 13731 1727203881.25854: WORKER PROCESS EXITING ok: [managed-node3] => {} MSG: +++++ Success in test 'Given two DHCP-enabled network interfaces, when creating a bond profile with them, then the controller device and bond port profiles are present and the specified bond options are set for the controller device.' +++++ 13731 1727203881.25903: no more pending results, returning what we have 13731 1727203881.25906: results queue empty 13731 1727203881.25907: checking for any_errors_fatal 13731 1727203881.25914: done checking for any_errors_fatal 13731 1727203881.25915: checking for max_fail_percentage 13731 1727203881.25916: done checking for max_fail_percentage 13731 1727203881.25917: checking to see if all hosts have failed and the running result is not ok 13731 1727203881.25918: done checking to see if all hosts have failed 13731 1727203881.25919: getting the remaining hosts for this loop 13731 1727203881.25920: done getting the remaining hosts for this loop 13731 1727203881.25924: getting the next task for host managed-node3 13731 1727203881.25931: done getting next task for host managed-node3 13731 1727203881.25934: ^ task is: TASK: Cleanup 13731 1727203881.25937: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203881.25943: getting variables 13731 1727203881.25944: in VariableManager get_vars() 13731 1727203881.25985: Calling all_inventory to load vars for managed-node3 13731 1727203881.25988: Calling groups_inventory to load vars for managed-node3 13731 1727203881.25990: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.25999: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.26002: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.26004: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.26770: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.27709: done with get_vars() 13731 1727203881.27723: done getting variables TASK [Cleanup] ***************************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:66 Tuesday 24 September 2024 14:51:21 -0400 (0:00:00.035) 0:00:59.508 ***** 13731 1727203881.27789: entering _queue_task() for managed-node3/include_tasks 13731 1727203881.28009: worker is 1 (out of 1 available) 13731 1727203881.28023: exiting _queue_task() for managed-node3/include_tasks 13731 1727203881.28036: done queuing things up, now waiting for results queue to drain 13731 1727203881.28038: waiting for pending results... 13731 1727203881.28220: running TaskExecutor() for managed-node3/TASK: Cleanup 13731 1727203881.28294: in run() - task 028d2410-947f-82dc-c122-0000000008f5 13731 1727203881.28307: variable 'ansible_search_path' from source: unknown 13731 1727203881.28310: variable 'ansible_search_path' from source: unknown 13731 1727203881.28343: variable 'lsr_cleanup' from source: include params 13731 1727203881.28488: variable 'lsr_cleanup' from source: include params 13731 1727203881.28541: variable 'omit' from source: magic vars 13731 1727203881.28644: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.28651: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.28660: variable 'omit' from source: magic vars 13731 1727203881.28828: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.28835: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.28841: variable 'item' from source: unknown 13731 1727203881.28891: variable 'item' from source: unknown 13731 1727203881.28915: variable 'item' from source: unknown 13731 1727203881.28957: variable 'item' from source: unknown 13731 1727203881.29078: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.29082: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.29085: variable 'omit' from source: magic vars 13731 1727203881.29161: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.29168: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.29173: variable 'item' from source: unknown 13731 1727203881.29219: variable 'item' from source: unknown 13731 1727203881.29238: variable 'item' from source: unknown 13731 1727203881.29282: variable 'item' from source: unknown 13731 1727203881.29340: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.29351: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.29354: variable 'omit' from source: magic vars 13731 1727203881.29462: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.29465: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.29468: variable 'item' from source: unknown 13731 1727203881.29511: variable 'item' from source: unknown 13731 1727203881.29530: variable 'item' from source: unknown 13731 1727203881.29575: variable 'item' from source: unknown 13731 1727203881.29630: dumping result to json 13731 1727203881.29633: done dumping result, returning 13731 1727203881.29635: done running TaskExecutor() for managed-node3/TASK: Cleanup [028d2410-947f-82dc-c122-0000000008f5] 13731 1727203881.29638: sending task result for task 028d2410-947f-82dc-c122-0000000008f5 13731 1727203881.29672: done sending task result for task 028d2410-947f-82dc-c122-0000000008f5 13731 1727203881.29675: WORKER PROCESS EXITING 13731 1727203881.29700: no more pending results, returning what we have 13731 1727203881.29706: in VariableManager get_vars() 13731 1727203881.29750: Calling all_inventory to load vars for managed-node3 13731 1727203881.29752: Calling groups_inventory to load vars for managed-node3 13731 1727203881.29755: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.29770: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.29772: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.29782: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.30552: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.31391: done with get_vars() 13731 1727203881.31407: variable 'ansible_search_path' from source: unknown 13731 1727203881.31408: variable 'ansible_search_path' from source: unknown 13731 1727203881.31434: variable 'ansible_search_path' from source: unknown 13731 1727203881.31435: variable 'ansible_search_path' from source: unknown 13731 1727203881.31452: variable 'ansible_search_path' from source: unknown 13731 1727203881.31453: variable 'ansible_search_path' from source: unknown 13731 1727203881.31470: we have included files to process 13731 1727203881.31471: generating all_blocks data 13731 1727203881.31472: done generating all_blocks data 13731 1727203881.31477: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml 13731 1727203881.31478: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml 13731 1727203881.31479: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml 13731 1727203881.31574: in VariableManager get_vars() 13731 1727203881.31593: done with get_vars() 13731 1727203881.31596: variable 'omit' from source: magic vars 13731 1727203881.31622: variable 'omit' from source: magic vars 13731 1727203881.31653: in VariableManager get_vars() 13731 1727203881.31664: done with get_vars() 13731 1727203881.31683: in VariableManager get_vars() 13731 1727203881.31696: done with get_vars() 13731 1727203881.31719: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/defaults/main.yml 13731 1727203881.31790: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/meta/main.yml 13731 1727203881.31870: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml 13731 1727203881.32093: in VariableManager get_vars() 13731 1727203881.32108: done with get_vars() redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13731 1727203881.33367: done processing included file 13731 1727203881.33369: iterating over new_blocks loaded from include file 13731 1727203881.33370: in VariableManager get_vars() 13731 1727203881.33386: done with get_vars() 13731 1727203881.33387: filtering new block on tags 13731 1727203881.33588: done filtering new block on tags 13731 1727203881.33591: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml for managed-node3 => (item=tasks/cleanup_bond_profile+device.yml) 13731 1727203881.33595: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml 13731 1727203881.33595: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml 13731 1727203881.33597: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml 13731 1727203881.33808: done processing included file 13731 1727203881.33810: iterating over new_blocks loaded from include file 13731 1727203881.33811: in VariableManager get_vars() 13731 1727203881.33824: done with get_vars() 13731 1727203881.33825: filtering new block on tags 13731 1727203881.33897: done filtering new block on tags 13731 1727203881.33899: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml for managed-node3 => (item=tasks/remove_test_interfaces_with_dhcp.yml) 13731 1727203881.33901: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml 13731 1727203881.33906: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml 13731 1727203881.33908: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml 13731 1727203881.34121: done processing included file 13731 1727203881.34123: iterating over new_blocks loaded from include file 13731 1727203881.34123: in VariableManager get_vars() 13731 1727203881.34138: done with get_vars() 13731 1727203881.34139: filtering new block on tags 13731 1727203881.34157: done filtering new block on tags 13731 1727203881.34158: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml for managed-node3 => (item=tasks/check_network_dns.yml) 13731 1727203881.34161: extending task lists for all hosts with included blocks 13731 1727203881.39456: done extending task lists 13731 1727203881.39457: done processing included files 13731 1727203881.39458: results queue empty 13731 1727203881.39458: checking for any_errors_fatal 13731 1727203881.39461: done checking for any_errors_fatal 13731 1727203881.39462: checking for max_fail_percentage 13731 1727203881.39462: done checking for max_fail_percentage 13731 1727203881.39463: checking to see if all hosts have failed and the running result is not ok 13731 1727203881.39464: done checking to see if all hosts have failed 13731 1727203881.39464: getting the remaining hosts for this loop 13731 1727203881.39465: done getting the remaining hosts for this loop 13731 1727203881.39467: getting the next task for host managed-node3 13731 1727203881.39470: done getting next task for host managed-node3 13731 1727203881.39472: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13731 1727203881.39474: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203881.39484: getting variables 13731 1727203881.39485: in VariableManager get_vars() 13731 1727203881.39499: Calling all_inventory to load vars for managed-node3 13731 1727203881.39500: Calling groups_inventory to load vars for managed-node3 13731 1727203881.39501: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.39505: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.39506: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.39508: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.40124: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.40957: done with get_vars() 13731 1727203881.40976: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:4 Tuesday 24 September 2024 14:51:21 -0400 (0:00:00.132) 0:00:59.641 ***** 13731 1727203881.41027: entering _queue_task() for managed-node3/include_tasks 13731 1727203881.41300: worker is 1 (out of 1 available) 13731 1727203881.41313: exiting _queue_task() for managed-node3/include_tasks 13731 1727203881.41326: done queuing things up, now waiting for results queue to drain 13731 1727203881.41328: waiting for pending results... 13731 1727203881.41515: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role 13731 1727203881.41616: in run() - task 028d2410-947f-82dc-c122-000000000e0c 13731 1727203881.41628: variable 'ansible_search_path' from source: unknown 13731 1727203881.41632: variable 'ansible_search_path' from source: unknown 13731 1727203881.41660: calling self._execute() 13731 1727203881.41738: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.41742: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.41751: variable 'omit' from source: magic vars 13731 1727203881.42035: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.42044: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.42050: _execute() done 13731 1727203881.42053: dumping result to json 13731 1727203881.42056: done dumping result, returning 13731 1727203881.42062: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role [028d2410-947f-82dc-c122-000000000e0c] 13731 1727203881.42070: sending task result for task 028d2410-947f-82dc-c122-000000000e0c 13731 1727203881.42157: done sending task result for task 028d2410-947f-82dc-c122-000000000e0c 13731 1727203881.42160: WORKER PROCESS EXITING 13731 1727203881.42209: no more pending results, returning what we have 13731 1727203881.42214: in VariableManager get_vars() 13731 1727203881.42267: Calling all_inventory to load vars for managed-node3 13731 1727203881.42278: Calling groups_inventory to load vars for managed-node3 13731 1727203881.42281: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.42293: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.42295: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.42298: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.43157: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.44022: done with get_vars() 13731 1727203881.44035: variable 'ansible_search_path' from source: unknown 13731 1727203881.44036: variable 'ansible_search_path' from source: unknown 13731 1727203881.44065: we have included files to process 13731 1727203881.44066: generating all_blocks data 13731 1727203881.44067: done generating all_blocks data 13731 1727203881.44069: processing included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203881.44069: loading included file: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203881.44071: Loading data from /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml 13731 1727203881.44450: done processing included file 13731 1727203881.44451: iterating over new_blocks loaded from include file 13731 1727203881.44452: in VariableManager get_vars() 13731 1727203881.44473: done with get_vars() 13731 1727203881.44477: filtering new block on tags 13731 1727203881.44498: done filtering new block on tags 13731 1727203881.44499: in VariableManager get_vars() 13731 1727203881.44515: done with get_vars() 13731 1727203881.44516: filtering new block on tags 13731 1727203881.44545: done filtering new block on tags 13731 1727203881.44547: in VariableManager get_vars() 13731 1727203881.44566: done with get_vars() 13731 1727203881.44567: filtering new block on tags 13731 1727203881.44596: done filtering new block on tags 13731 1727203881.44598: done iterating over new_blocks loaded from include file included: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml for managed-node3 13731 1727203881.44601: extending task lists for all hosts with included blocks 13731 1727203881.45678: done extending task lists 13731 1727203881.45680: done processing included files 13731 1727203881.45681: results queue empty 13731 1727203881.45682: checking for any_errors_fatal 13731 1727203881.45686: done checking for any_errors_fatal 13731 1727203881.45687: checking for max_fail_percentage 13731 1727203881.45688: done checking for max_fail_percentage 13731 1727203881.45689: checking to see if all hosts have failed and the running result is not ok 13731 1727203881.45690: done checking to see if all hosts have failed 13731 1727203881.45690: getting the remaining hosts for this loop 13731 1727203881.45692: done getting the remaining hosts for this loop 13731 1727203881.45694: getting the next task for host managed-node3 13731 1727203881.45700: done getting next task for host managed-node3 13731 1727203881.45703: ^ task is: TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13731 1727203881.45707: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203881.45717: getting variables 13731 1727203881.45718: in VariableManager get_vars() 13731 1727203881.45736: Calling all_inventory to load vars for managed-node3 13731 1727203881.45739: Calling groups_inventory to load vars for managed-node3 13731 1727203881.45740: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.45746: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.45748: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.45751: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.46860: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.48371: done with get_vars() 13731 1727203881.48397: done getting variables TASK [fedora.linux_system_roles.network : Ensure ansible_facts used by role are present] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:3 Tuesday 24 September 2024 14:51:21 -0400 (0:00:00.074) 0:00:59.715 ***** 13731 1727203881.48484: entering _queue_task() for managed-node3/setup 13731 1727203881.48788: worker is 1 (out of 1 available) 13731 1727203881.48803: exiting _queue_task() for managed-node3/setup 13731 1727203881.48815: done queuing things up, now waiting for results queue to drain 13731 1727203881.48816: waiting for pending results... 13731 1727203881.49296: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present 13731 1727203881.49305: in run() - task 028d2410-947f-82dc-c122-000000000fe0 13731 1727203881.49308: variable 'ansible_search_path' from source: unknown 13731 1727203881.49317: variable 'ansible_search_path' from source: unknown 13731 1727203881.49359: calling self._execute() 13731 1727203881.49468: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.49485: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.49501: variable 'omit' from source: magic vars 13731 1727203881.49911: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.49930: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.50159: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203881.51929: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203881.51982: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203881.52011: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203881.52040: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203881.52060: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203881.52118: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203881.52141: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203881.52159: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203881.52188: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203881.52198: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203881.52240: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203881.52254: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203881.52272: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203881.52298: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203881.52308: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203881.52415: variable '__network_required_facts' from source: role '' defaults 13731 1727203881.52422: variable 'ansible_facts' from source: unknown 13731 1727203881.52838: Evaluated conditional (__network_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 13731 1727203881.52841: when evaluation is False, skipping this task 13731 1727203881.52844: _execute() done 13731 1727203881.52847: dumping result to json 13731 1727203881.52849: done dumping result, returning 13731 1727203881.52854: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure ansible_facts used by role are present [028d2410-947f-82dc-c122-000000000fe0] 13731 1727203881.52859: sending task result for task 028d2410-947f-82dc-c122-000000000fe0 13731 1727203881.52943: done sending task result for task 028d2410-947f-82dc-c122-000000000fe0 13731 1727203881.52946: WORKER PROCESS EXITING skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203881.52992: no more pending results, returning what we have 13731 1727203881.52996: results queue empty 13731 1727203881.52996: checking for any_errors_fatal 13731 1727203881.52998: done checking for any_errors_fatal 13731 1727203881.52999: checking for max_fail_percentage 13731 1727203881.53000: done checking for max_fail_percentage 13731 1727203881.53001: checking to see if all hosts have failed and the running result is not ok 13731 1727203881.53002: done checking to see if all hosts have failed 13731 1727203881.53003: getting the remaining hosts for this loop 13731 1727203881.53004: done getting the remaining hosts for this loop 13731 1727203881.53007: getting the next task for host managed-node3 13731 1727203881.53018: done getting next task for host managed-node3 13731 1727203881.53022: ^ task is: TASK: fedora.linux_system_roles.network : Check if system is ostree 13731 1727203881.53027: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203881.53047: getting variables 13731 1727203881.53049: in VariableManager get_vars() 13731 1727203881.53101: Calling all_inventory to load vars for managed-node3 13731 1727203881.53104: Calling groups_inventory to load vars for managed-node3 13731 1727203881.53106: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.53115: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.53118: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.53126: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.53986: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.54871: done with get_vars() 13731 1727203881.54890: done getting variables TASK [fedora.linux_system_roles.network : Check if system is ostree] *********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:12 Tuesday 24 September 2024 14:51:21 -0400 (0:00:00.064) 0:00:59.780 ***** 13731 1727203881.54963: entering _queue_task() for managed-node3/stat 13731 1727203881.55203: worker is 1 (out of 1 available) 13731 1727203881.55216: exiting _queue_task() for managed-node3/stat 13731 1727203881.55227: done queuing things up, now waiting for results queue to drain 13731 1727203881.55229: waiting for pending results... 13731 1727203881.55414: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if system is ostree 13731 1727203881.55525: in run() - task 028d2410-947f-82dc-c122-000000000fe2 13731 1727203881.55538: variable 'ansible_search_path' from source: unknown 13731 1727203881.55541: variable 'ansible_search_path' from source: unknown 13731 1727203881.55570: calling self._execute() 13731 1727203881.55644: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.55648: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.55656: variable 'omit' from source: magic vars 13731 1727203881.55932: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.55941: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.56053: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203881.56247: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203881.56286: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203881.56310: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203881.56340: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203881.56431: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203881.56452: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203881.56471: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203881.56490: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203881.56555: variable '__network_is_ostree' from source: set_fact 13731 1727203881.56559: Evaluated conditional (not __network_is_ostree is defined): False 13731 1727203881.56565: when evaluation is False, skipping this task 13731 1727203881.56567: _execute() done 13731 1727203881.56570: dumping result to json 13731 1727203881.56572: done dumping result, returning 13731 1727203881.56579: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if system is ostree [028d2410-947f-82dc-c122-000000000fe2] 13731 1727203881.56585: sending task result for task 028d2410-947f-82dc-c122-000000000fe2 13731 1727203881.56669: done sending task result for task 028d2410-947f-82dc-c122-000000000fe2 13731 1727203881.56672: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13731 1727203881.56722: no more pending results, returning what we have 13731 1727203881.56728: results queue empty 13731 1727203881.56729: checking for any_errors_fatal 13731 1727203881.56737: done checking for any_errors_fatal 13731 1727203881.56738: checking for max_fail_percentage 13731 1727203881.56740: done checking for max_fail_percentage 13731 1727203881.56741: checking to see if all hosts have failed and the running result is not ok 13731 1727203881.56742: done checking to see if all hosts have failed 13731 1727203881.56742: getting the remaining hosts for this loop 13731 1727203881.56744: done getting the remaining hosts for this loop 13731 1727203881.56747: getting the next task for host managed-node3 13731 1727203881.56754: done getting next task for host managed-node3 13731 1727203881.56758: ^ task is: TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13731 1727203881.56766: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203881.56787: getting variables 13731 1727203881.56789: in VariableManager get_vars() 13731 1727203881.56827: Calling all_inventory to load vars for managed-node3 13731 1727203881.56830: Calling groups_inventory to load vars for managed-node3 13731 1727203881.56832: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.56839: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.56841: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.56844: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.57609: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.58584: done with get_vars() 13731 1727203881.58599: done getting variables 13731 1727203881.58642: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Set flag to indicate system is ostree] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:17 Tuesday 24 September 2024 14:51:21 -0400 (0:00:00.037) 0:00:59.817 ***** 13731 1727203881.58670: entering _queue_task() for managed-node3/set_fact 13731 1727203881.58896: worker is 1 (out of 1 available) 13731 1727203881.58907: exiting _queue_task() for managed-node3/set_fact 13731 1727203881.58919: done queuing things up, now waiting for results queue to drain 13731 1727203881.58921: waiting for pending results... 13731 1727203881.59111: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree 13731 1727203881.59224: in run() - task 028d2410-947f-82dc-c122-000000000fe3 13731 1727203881.59236: variable 'ansible_search_path' from source: unknown 13731 1727203881.59239: variable 'ansible_search_path' from source: unknown 13731 1727203881.59269: calling self._execute() 13731 1727203881.59345: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.59348: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.59357: variable 'omit' from source: magic vars 13731 1727203881.59631: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.59640: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.59751: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203881.59943: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203881.59978: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203881.60002: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203881.60031: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203881.60122: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203881.60144: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203881.60164: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203881.60182: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203881.60243: variable '__network_is_ostree' from source: set_fact 13731 1727203881.60255: Evaluated conditional (not __network_is_ostree is defined): False 13731 1727203881.60258: when evaluation is False, skipping this task 13731 1727203881.60260: _execute() done 13731 1727203881.60263: dumping result to json 13731 1727203881.60269: done dumping result, returning 13731 1727203881.60277: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Set flag to indicate system is ostree [028d2410-947f-82dc-c122-000000000fe3] 13731 1727203881.60281: sending task result for task 028d2410-947f-82dc-c122-000000000fe3 13731 1727203881.60356: done sending task result for task 028d2410-947f-82dc-c122-000000000fe3 13731 1727203881.60359: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "not __network_is_ostree is defined", "skip_reason": "Conditional result was False" } 13731 1727203881.60403: no more pending results, returning what we have 13731 1727203881.60407: results queue empty 13731 1727203881.60408: checking for any_errors_fatal 13731 1727203881.60415: done checking for any_errors_fatal 13731 1727203881.60416: checking for max_fail_percentage 13731 1727203881.60418: done checking for max_fail_percentage 13731 1727203881.60419: checking to see if all hosts have failed and the running result is not ok 13731 1727203881.60420: done checking to see if all hosts have failed 13731 1727203881.60420: getting the remaining hosts for this loop 13731 1727203881.60422: done getting the remaining hosts for this loop 13731 1727203881.60425: getting the next task for host managed-node3 13731 1727203881.60436: done getting next task for host managed-node3 13731 1727203881.60440: ^ task is: TASK: fedora.linux_system_roles.network : Check which services are running 13731 1727203881.60445: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203881.60464: getting variables 13731 1727203881.60465: in VariableManager get_vars() 13731 1727203881.60505: Calling all_inventory to load vars for managed-node3 13731 1727203881.60509: Calling groups_inventory to load vars for managed-node3 13731 1727203881.60511: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203881.60518: Calling all_plugins_play to load vars for managed-node3 13731 1727203881.60521: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203881.60523: Calling groups_plugins_play to load vars for managed-node3 13731 1727203881.61266: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203881.62132: done with get_vars() 13731 1727203881.62147: done getting variables TASK [fedora.linux_system_roles.network : Check which services are running] **** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 Tuesday 24 September 2024 14:51:21 -0400 (0:00:00.035) 0:00:59.853 ***** 13731 1727203881.62217: entering _queue_task() for managed-node3/service_facts 13731 1727203881.62431: worker is 1 (out of 1 available) 13731 1727203881.62444: exiting _queue_task() for managed-node3/service_facts 13731 1727203881.62456: done queuing things up, now waiting for results queue to drain 13731 1727203881.62457: waiting for pending results... 13731 1727203881.62632: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which services are running 13731 1727203881.62740: in run() - task 028d2410-947f-82dc-c122-000000000fe5 13731 1727203881.62751: variable 'ansible_search_path' from source: unknown 13731 1727203881.62756: variable 'ansible_search_path' from source: unknown 13731 1727203881.62787: calling self._execute() 13731 1727203881.62860: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.62867: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.62877: variable 'omit' from source: magic vars 13731 1727203881.63150: variable 'ansible_distribution_major_version' from source: facts 13731 1727203881.63159: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203881.63169: variable 'omit' from source: magic vars 13731 1727203881.63235: variable 'omit' from source: magic vars 13731 1727203881.63256: variable 'omit' from source: magic vars 13731 1727203881.63290: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203881.63315: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203881.63330: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203881.63346: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203881.63356: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203881.63383: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203881.63386: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.63389: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.63455: Set connection var ansible_pipelining to False 13731 1727203881.63459: Set connection var ansible_shell_type to sh 13731 1727203881.63468: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203881.63472: Set connection var ansible_connection to ssh 13731 1727203881.63478: Set connection var ansible_shell_executable to /bin/sh 13731 1727203881.63484: Set connection var ansible_timeout to 10 13731 1727203881.63500: variable 'ansible_shell_executable' from source: unknown 13731 1727203881.63503: variable 'ansible_connection' from source: unknown 13731 1727203881.63506: variable 'ansible_module_compression' from source: unknown 13731 1727203881.63509: variable 'ansible_shell_type' from source: unknown 13731 1727203881.63511: variable 'ansible_shell_executable' from source: unknown 13731 1727203881.63514: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203881.63516: variable 'ansible_pipelining' from source: unknown 13731 1727203881.63519: variable 'ansible_timeout' from source: unknown 13731 1727203881.63523: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203881.63672: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203881.63679: variable 'omit' from source: magic vars 13731 1727203881.63684: starting attempt loop 13731 1727203881.63687: running the handler 13731 1727203881.63699: _low_level_execute_command(): starting 13731 1727203881.63706: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203881.64216: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203881.64220: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.64223: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203881.64225: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.64280: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203881.64283: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203881.64290: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203881.64331: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203881.66287: stdout chunk (state=3): >>>/root <<< 13731 1727203881.66387: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203881.66416: stderr chunk (state=3): >>><<< 13731 1727203881.66419: stdout chunk (state=3): >>><<< 13731 1727203881.66437: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203881.66449: _low_level_execute_command(): starting 13731 1727203881.66455: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317 `" && echo ansible-tmp-1727203881.6643684-18118-59095570388317="` echo /root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317 `" ) && sleep 0' 13731 1727203881.66896: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203881.66899: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203881.66902: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.66911: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203881.66913: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203881.66916: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.66958: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203881.66961: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203881.66966: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203881.67001: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203881.68884: stdout chunk (state=3): >>>ansible-tmp-1727203881.6643684-18118-59095570388317=/root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317 <<< 13731 1727203881.68992: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203881.69017: stderr chunk (state=3): >>><<< 13731 1727203881.69020: stdout chunk (state=3): >>><<< 13731 1727203881.69034: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203881.6643684-18118-59095570388317=/root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203881.69071: variable 'ansible_module_compression' from source: unknown 13731 1727203881.69111: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.service_facts-ZIP_DEFLATED 13731 1727203881.69141: variable 'ansible_facts' from source: unknown 13731 1727203881.69201: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/AnsiballZ_service_facts.py 13731 1727203881.69296: Sending initial data 13731 1727203881.69299: Sent initial data (161 bytes) 13731 1727203881.69733: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203881.69738: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203881.69741: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203881.69744: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203881.69746: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.69802: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203881.69805: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203881.69834: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203881.71602: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 <<< 13731 1727203881.71606: stderr chunk (state=3): >>>debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203881.71634: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203881.71668: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmptixrrz17 /root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/AnsiballZ_service_facts.py <<< 13731 1727203881.71671: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/AnsiballZ_service_facts.py" <<< 13731 1727203881.71700: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmptixrrz17" to remote "/root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/AnsiballZ_service_facts.py" <<< 13731 1727203881.71707: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/AnsiballZ_service_facts.py" <<< 13731 1727203881.72206: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203881.72246: stderr chunk (state=3): >>><<< 13731 1727203881.72249: stdout chunk (state=3): >>><<< 13731 1727203881.72305: done transferring module to remote 13731 1727203881.72313: _low_level_execute_command(): starting 13731 1727203881.72318: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/ /root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/AnsiballZ_service_facts.py && sleep 0' 13731 1727203881.72731: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203881.72735: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.72747: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.72804: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203881.72808: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203881.72846: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203881.74593: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203881.74618: stderr chunk (state=3): >>><<< 13731 1727203881.74621: stdout chunk (state=3): >>><<< 13731 1727203881.74632: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203881.74635: _low_level_execute_command(): starting 13731 1727203881.74640: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/AnsiballZ_service_facts.py && sleep 0' 13731 1727203881.75035: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203881.75039: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.75058: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203881.75105: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203881.75108: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203881.75157: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203883.27119: stdout chunk (state=3): >>> {"ansible_facts": {"services": {"audit-rules.service": {"name": "audit-rules.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dm-event.service": {"name": "dm-event.service", "state": "stopped", "status": "static", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-lvmpolld.service": {"name": "lvm2-lvmpolld.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-monitor.service": {"name": "lvm2-monitor.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@dm_mod.service": {"name": "modprobe@dm_mod.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@efi_pstore.service": {"name": "modprobe@efi_pstore.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@loop.service": {"name": "modprobe@loop.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "running", "status": "enabled", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "pcscd.service": {"name": "pcscd.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ssh-host-keys-migration.service": {"name": "ssh-host-keys-migration.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-battery-check.service": {"name": "systemd-battery-check.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-confext.service": {"name": "systemd-confext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-clear.service": {"name": "systemd-hibernate-clear.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-resume.service": {"name": "systemd-hibernate-resume.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-oomd.service": {"name": "systemd-oomd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-quotacheck-root.service": {"name": "systemd-quotacheck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-soft-reboot.service": {"name": "systemd-soft-reboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev-early.service": {"name": "systemd-tmpfiles-setup-dev-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup-early.service": {"name": "systemd-tpm2-setup-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup.service": {"name": "systemd-tpm2-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-load-credentials.service": {"name": "systemd-udev-load-credentials.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "blk-availability.service": {"name": "blk-availability.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "capsule@.service": {"name": "capsule@.service", "state": "unknown", "status": "static", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd.service": {"name": "dhcpcd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd@.service": {"name": "dhcpcd@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fips-crypto-policy-overlay.service": {"name": "fips-crypto-policy-overlay.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fsidd.service": {"name": "fsidd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "lvm-devices-import.service": {"name": "lvm-devices-import.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "polkit.service": {"name": "polkit.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon-root.service": {"name": "quotaon-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "quotaon@.service": {"name": "quotaon@.service", "state": "unknown", "status": "static", "source": "systemd"}, "rpmdb-migrate.service": {"name": "rpmdb-migrate.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-bootctl@.service": {"name": "systemd-bootctl@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-creds@.service": {"name": "systemd-creds@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald-sync@.service": {"name": "systemd-journald-sync@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrextend@.service": {"name": "systemd-pcrextend@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrlock-file-system.service": {"name": "systemd-pcrlock-file-system.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-code.service": {"name": "systemd-pcrlock-firmware-code.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-config.service": {"name": "systemd-pcrlock-firmware-config.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-machine-id.service": {"name": "systemd-pcrlock-machine-id.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-make-policy.service": {"name": "systemd-pcrlock-make-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-authority.service": {"name": "systemd-pcrlock-secureboot-authority.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-policy.service": {"name": "systemd-pcrlock-secureboot-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock@.service": {"name": "systemd-pcrlock@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-quotacheck@.service": {"name": "systemd-quotacheck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysext@.service": {"name": "systemd-sysext@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} <<< 13731 1727203883.28682: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203883.28687: stdout chunk (state=3): >>><<< 13731 1727203883.28689: stderr chunk (state=3): >>><<< 13731 1727203883.28693: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"services": {"audit-rules.service": {"name": "audit-rules.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "auditd.service": {"name": "auditd.service", "state": "running", "status": "enabled", "source": "systemd"}, "auth-rpcgss-module.service": {"name": "auth-rpcgss-module.service", "state": "stopped", "status": "static", "source": "systemd"}, "autofs.service": {"name": "autofs.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "chronyd.service": {"name": "chronyd.service", "state": "running", "status": "enabled", "source": "systemd"}, "cloud-config.service": {"name": "cloud-config.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-final.service": {"name": "cloud-final.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init-local.service": {"name": "cloud-init-local.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "cloud-init.service": {"name": "cloud-init.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "crond.service": {"name": "crond.service", "state": "running", "status": "enabled", "source": "systemd"}, "dbus-broker.service": {"name": "dbus-broker.service", "state": "running", "status": "enabled", "source": "systemd"}, "display-manager.service": {"name": "display-manager.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "dm-event.service": {"name": "dm-event.service", "state": "stopped", "status": "static", "source": "systemd"}, "dnf-makecache.service": {"name": "dnf-makecache.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-cmdline.service": {"name": "dracut-cmdline.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-initqueue.service": {"name": "dracut-initqueue.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-mount.service": {"name": "dracut-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-mount.service": {"name": "dracut-pre-mount.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-pivot.service": {"name": "dracut-pre-pivot.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-trigger.service": {"name": "dracut-pre-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-pre-udev.service": {"name": "dracut-pre-udev.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown-onfailure.service": {"name": "dracut-shutdown-onfailure.service", "state": "stopped", "status": "static", "source": "systemd"}, "dracut-shutdown.service": {"name": "dracut-shutdown.service", "state": "stopped", "status": "static", "source": "systemd"}, "emergency.service": {"name": "emergency.service", "state": "stopped", "status": "static", "source": "systemd"}, "fstrim.service": {"name": "fstrim.service", "state": "stopped", "status": "static", "source": "systemd"}, "getty@tty1.service": {"name": "getty@tty1.service", "state": "running", "status": "active", "source": "systemd"}, "gssproxy.service": {"name": "gssproxy.service", "state": "running", "status": "disabled", "source": "systemd"}, "hv_kvp_daemon.service": {"name": "hv_kvp_daemon.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "initrd-cleanup.service": {"name": "initrd-cleanup.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-parse-etc.service": {"name": "initrd-parse-etc.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-switch-root.service": {"name": "initrd-switch-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "initrd-udevadm-cleanup-db.service": {"name": "initrd-udevadm-cleanup-db.service", "state": "stopped", "status": "static", "source": "systemd"}, "irqbalance.service": {"name": "irqbalance.service", "state": "running", "status": "enabled", "source": "systemd"}, "kdump.service": {"name": "kdump.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "kmod-static-nodes.service": {"name": "kmod-static-nodes.service", "state": "stopped", "status": "static", "source": "systemd"}, "ldconfig.service": {"name": "ldconfig.service", "state": "stopped", "status": "static", "source": "systemd"}, "logrotate.service": {"name": "logrotate.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-lvmpolld.service": {"name": "lvm2-lvmpolld.service", "state": "stopped", "status": "static", "source": "systemd"}, "lvm2-monitor.service": {"name": "lvm2-monitor.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "modprobe@configfs.service": {"name": "modprobe@configfs.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@dm_mod.service": {"name": "modprobe@dm_mod.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@drm.service": {"name": "modprobe@drm.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@efi_pstore.service": {"name": "modprobe@efi_pstore.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@fuse.service": {"name": "modprobe@fuse.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "modprobe@loop.service": {"name": "modprobe@loop.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "network.service": {"name": "network.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "NetworkManager-dispatcher.service": {"name": "NetworkManager-dispatcher.service", "state": "running", "status": "enabled", "source": "systemd"}, "NetworkManager-wait-online.service": {"name": "NetworkManager-wait-online.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "NetworkManager.service": {"name": "NetworkManager.service", "state": "running", "status": "enabled", "source": "systemd"}, "nfs-idmapd.service": {"name": "nfs-idmapd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-mountd.service": {"name": "nfs-mountd.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfs-server.service": {"name": "nfs-server.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "nfs-utils.service": {"name": "nfs-utils.service", "state": "stopped", "status": "static", "source": "systemd"}, "nfsdcld.service": {"name": "nfsdcld.service", "state": "stopped", "status": "static", "source": "systemd"}, "ntpd.service": {"name": "ntpd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ntpdate.service": {"name": "ntpdate.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "pcscd.service": {"name": "pcscd.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "plymouth-quit-wait.service": {"name": "plymouth-quit-wait.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "plymouth-start.service": {"name": "plymouth-start.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rc-local.service": {"name": "rc-local.service", "state": "stopped", "status": "static", "source": "systemd"}, "rescue.service": {"name": "rescue.service", "state": "stopped", "status": "static", "source": "systemd"}, "restraintd.service": {"name": "restraintd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rngd.service": {"name": "rngd.service", "state": "running", "status": "enabled", "source": "systemd"}, "rpc-gssd.service": {"name": "rpc-gssd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd-notify.service": {"name": "rpc-statd-notify.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-statd.service": {"name": "rpc-statd.service", "state": "stopped", "status": "static", "source": "systemd"}, "rpc-svcgssd.service": {"name": "rpc-svcgssd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "rpcbind.service": {"name": "rpcbind.service", "state": "running", "status": "enabled", "source": "systemd"}, "rsyslog.service": {"name": "rsyslog.service", "state": "running", "status": "enabled", "source": "systemd"}, "selinux-autorelabel-mark.service": {"name": "selinux-autorelabel-mark.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "serial-getty@ttyS0.service": {"name": "serial-getty@ttyS0.service", "state": "running", "status": "active", "source": "systemd"}, "sntp.service": {"name": "sntp.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "ssh-host-keys-migration.service": {"name": "ssh-host-keys-migration.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "sshd-keygen.service": {"name": "sshd-keygen.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "sshd-keygen@ecdsa.service": {"name": "sshd-keygen@ecdsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@ed25519.service": {"name": "sshd-keygen@ed25519.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd-keygen@rsa.service": {"name": "sshd-keygen@rsa.service", "state": "stopped", "status": "inactive", "source": "systemd"}, "sshd.service": {"name": "sshd.service", "state": "running", "status": "enabled", "source": "systemd"}, "sssd-kcm.service": {"name": "sssd-kcm.service", "state": "stopped", "status": "indirect", "source": "systemd"}, "sssd.service": {"name": "sssd.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "syslog.service": {"name": "syslog.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-ask-password-console.service": {"name": "systemd-ask-password-console.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-ask-password-wall.service": {"name": "systemd-ask-password-wall.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-battery-check.service": {"name": "systemd-battery-check.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-binfmt.service": {"name": "systemd-binfmt.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-boot-random-seed.service": {"name": "systemd-boot-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-confext.service": {"name": "systemd-confext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-firstboot.service": {"name": "systemd-firstboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-fsck-root.service": {"name": "systemd-fsck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-clear.service": {"name": "systemd-hibernate-clear.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hibernate-resume.service": {"name": "systemd-hibernate-resume.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hostnamed.service": {"name": "systemd-hostnamed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-hwdb-update.service": {"name": "systemd-hwdb-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-initctl.service": {"name": "systemd-initctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-catalog-update.service": {"name": "systemd-journal-catalog-update.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journal-flush.service": {"name": "systemd-journal-flush.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-journald.service": {"name": "systemd-journald.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-logind.service": {"name": "systemd-logind.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-machine-id-commit.service": {"name": "systemd-machine-id-commit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-modules-load.service": {"name": "systemd-modules-load.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-network-generator.service": {"name": "systemd-network-generator.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-networkd-wait-online.service": {"name": "systemd-networkd-wait-online.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-oomd.service": {"name": "systemd-oomd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-pcrmachine.service": {"name": "systemd-pcrmachine.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-initrd.service": {"name": "systemd-pcrphase-initrd.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase-sysinit.service": {"name": "systemd-pcrphase-sysinit.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pcrphase.service": {"name": "systemd-pcrphase.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-pstore.service": {"name": "systemd-pstore.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-quotacheck-root.service": {"name": "systemd-quotacheck-root.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-random-seed.service": {"name": "systemd-random-seed.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-remount-fs.service": {"name": "systemd-remount-fs.service", "state": "stopped", "status": "enabled-runtime", "source": "systemd"}, "systemd-repart.service": {"name": "systemd-repart.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-rfkill.service": {"name": "systemd-rfkill.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-soft-reboot.service": {"name": "systemd-soft-reboot.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysctl.service": {"name": "systemd-sysctl.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-sysext.service": {"name": "systemd-sysext.service", "state": "stopped", "status": "enabled", "source": "systemd"}, "systemd-sysusers.service": {"name": "systemd-sysusers.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-timesyncd.service": {"name": "systemd-timesyncd.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "systemd-tmpfiles-clean.service": {"name": "systemd-tmpfiles-clean.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev-early.service": {"name": "systemd-tmpfiles-setup-dev-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup-dev.service": {"name": "systemd-tmpfiles-setup-dev.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tmpfiles-setup.service": {"name": "systemd-tmpfiles-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup-early.service": {"name": "systemd-tpm2-setup-early.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-tpm2-setup.service": {"name": "systemd-tpm2-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-load-credentials.service": {"name": "systemd-udev-load-credentials.service", "state": "stopped", "status": "disabled", "source": "systemd"}, "systemd-udev-settle.service": {"name": "systemd-udev-settle.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udev-trigger.service": {"name": "systemd-udev-trigger.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-udevd.service": {"name": "systemd-udevd.service", "state": "running", "status": "static", "source": "systemd"}, "systemd-update-done.service": {"name": "systemd-update-done.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp-runlevel.service": {"name": "systemd-update-utmp-runlevel.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-update-utmp.service": {"name": "systemd-update-utmp.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-user-sessions.service": {"name": "systemd-user-sessions.service", "state": "stopped", "status": "static", "source": "systemd"}, "systemd-vconsole-setup.service": {"name": "systemd-vconsole-setup.service", "state": "stopped", "status": "static", "source": "systemd"}, "user-runtime-dir@0.service": {"name": "user-runtime-dir@0.service", "state": "stopped", "status": "active", "source": "systemd"}, "user@0.service": {"name": "user@0.service", "state": "running", "status": "active", "source": "systemd"}, "ypbind.service": {"name": "ypbind.service", "state": "stopped", "status": "not-found", "source": "systemd"}, "autovt@.service": {"name": "autovt@.service", "state": "unknown", "status": "alias", "source": "systemd"}, "blk-availability.service": {"name": "blk-availability.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "capsule@.service": {"name": "capsule@.service", "state": "unknown", "status": "static", "source": "systemd"}, "chrony-wait.service": {"name": "chrony-wait.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "chronyd-restricted.service": {"name": "chronyd-restricted.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "cloud-init-hotplugd.service": {"name": "cloud-init-hotplugd.service", "state": "inactive", "status": "static", "source": "systemd"}, "console-getty.service": {"name": "console-getty.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "container-getty@.service": {"name": "container-getty@.service", "state": "unknown", "status": "static", "source": "systemd"}, "dbus-org.freedesktop.hostname1.service": {"name": "dbus-org.freedesktop.hostname1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.locale1.service": {"name": "dbus-org.freedesktop.locale1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.login1.service": {"name": "dbus-org.freedesktop.login1.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.nm-dispatcher.service": {"name": "dbus-org.freedesktop.nm-dispatcher.service", "state": "active", "status": "alias", "source": "systemd"}, "dbus-org.freedesktop.timedate1.service": {"name": "dbus-org.freedesktop.timedate1.service", "state": "inactive", "status": "alias", "source": "systemd"}, "dbus.service": {"name": "dbus.service", "state": "active", "status": "alias", "source": "systemd"}, "debug-shell.service": {"name": "debug-shell.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd.service": {"name": "dhcpcd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dhcpcd@.service": {"name": "dhcpcd@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "dnf-system-upgrade-cleanup.service": {"name": "dnf-system-upgrade-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "dnf-system-upgrade.service": {"name": "dnf-system-upgrade.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "dnsmasq.service": {"name": "dnsmasq.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fips-crypto-policy-overlay.service": {"name": "fips-crypto-policy-overlay.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "firewalld.service": {"name": "firewalld.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "fsidd.service": {"name": "fsidd.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "getty@.service": {"name": "getty@.service", "state": "unknown", "status": "enabled", "source": "systemd"}, "grub-boot-indeterminate.service": {"name": "grub-boot-indeterminate.service", "state": "inactive", "status": "static", "source": "systemd"}, "grub2-systemd-integration.service": {"name": "grub2-systemd-integration.service", "state": "inactive", "status": "static", "source": "systemd"}, "kvm_stat.service": {"name": "kvm_stat.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "lvm-devices-import.service": {"name": "lvm-devices-import.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "man-db-cache-update.service": {"name": "man-db-cache-update.service", "state": "inactive", "status": "static", "source": "systemd"}, "man-db-restart-cache-update.service": {"name": "man-db-restart-cache-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "microcode.service": {"name": "microcode.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "modprobe@.service": {"name": "modprobe@.service", "state": "unknown", "status": "static", "source": "systemd"}, "nfs-blkmap.service": {"name": "nfs-blkmap.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nftables.service": {"name": "nftables.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nis-domainname.service": {"name": "nis-domainname.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "nm-priv-helper.service": {"name": "nm-priv-helper.service", "state": "inactive", "status": "static", "source": "systemd"}, "pam_namespace.service": {"name": "pam_namespace.service", "state": "inactive", "status": "static", "source": "systemd"}, "polkit.service": {"name": "polkit.service", "state": "inactive", "status": "static", "source": "systemd"}, "qemu-guest-agent.service": {"name": "qemu-guest-agent.service", "state": "inactive", "status": "enabled", "source": "systemd"}, "quotaon-root.service": {"name": "quotaon-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "quotaon@.service": {"name": "quotaon@.service", "state": "unknown", "status": "static", "source": "systemd"}, "rpmdb-migrate.service": {"name": "rpmdb-migrate.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "rpmdb-rebuild.service": {"name": "rpmdb-rebuild.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "selinux-autorelabel.service": {"name": "selinux-autorelabel.service", "state": "inactive", "status": "static", "source": "systemd"}, "selinux-check-proper-disable.service": {"name": "selinux-check-proper-disable.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "serial-getty@.service": {"name": "serial-getty@.service", "state": "unknown", "status": "indirect", "source": "systemd"}, "sshd-keygen@.service": {"name": "sshd-keygen@.service", "state": "unknown", "status": "disabled", "source": "systemd"}, "sshd@.service": {"name": "sshd@.service", "state": "unknown", "status": "static", "source": "systemd"}, "sssd-autofs.service": {"name": "sssd-autofs.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-nss.service": {"name": "sssd-nss.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pac.service": {"name": "sssd-pac.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-pam.service": {"name": "sssd-pam.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-ssh.service": {"name": "sssd-ssh.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "sssd-sudo.service": {"name": "sssd-sudo.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "system-update-cleanup.service": {"name": "system-update-cleanup.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-backlight@.service": {"name": "systemd-backlight@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-bless-boot.service": {"name": "systemd-bless-boot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-boot-check-no-failures.service": {"name": "systemd-boot-check-no-failures.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-boot-update.service": {"name": "systemd-boot-update.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-bootctl@.service": {"name": "systemd-bootctl@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-coredump@.service": {"name": "systemd-coredump@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-creds@.service": {"name": "systemd-creds@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-exit.service": {"name": "systemd-exit.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-fsck@.service": {"name": "systemd-fsck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-growfs-root.service": {"name": "systemd-growfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-growfs@.service": {"name": "systemd-growfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-halt.service": {"name": "systemd-halt.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hibernate.service": {"name": "systemd-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-hybrid-sleep.service": {"name": "systemd-hybrid-sleep.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-journald-sync@.service": {"name": "systemd-journald-sync@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-journald@.service": {"name": "systemd-journald@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-kexec.service": {"name": "systemd-kexec.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-localed.service": {"name": "systemd-localed.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrextend@.service": {"name": "systemd-pcrextend@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrfs-root.service": {"name": "systemd-pcrfs-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-pcrfs@.service": {"name": "systemd-pcrfs@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-pcrlock-file-system.service": {"name": "systemd-pcrlock-file-system.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-code.service": {"name": "systemd-pcrlock-firmware-code.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-firmware-config.service": {"name": "systemd-pcrlock-firmware-config.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-machine-id.service": {"name": "systemd-pcrlock-machine-id.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-make-policy.service": {"name": "systemd-pcrlock-make-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-authority.service": {"name": "systemd-pcrlock-secureboot-authority.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock-secureboot-policy.service": {"name": "systemd-pcrlock-secureboot-policy.service", "state": "inactive", "status": "disabled", "source": "systemd"}, "systemd-pcrlock@.service": {"name": "systemd-pcrlock@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-poweroff.service": {"name": "systemd-poweroff.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-quotacheck@.service": {"name": "systemd-quotacheck@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-reboot.service": {"name": "systemd-reboot.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend-then-hibernate.service": {"name": "systemd-suspend-then-hibernate.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-suspend.service": {"name": "systemd-suspend.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-sysext@.service": {"name": "systemd-sysext@.service", "state": "unknown", "status": "static", "source": "systemd"}, "systemd-sysupdate-reboot.service": {"name": "systemd-sysupdate-reboot.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-sysupdate.service": {"name": "systemd-sysupdate.service", "state": "inactive", "status": "indirect", "source": "systemd"}, "systemd-timedated.service": {"name": "systemd-timedated.service", "state": "inactive", "status": "static", "source": "systemd"}, "systemd-volatile-root.service": {"name": "systemd-volatile-root.service", "state": "inactive", "status": "static", "source": "systemd"}, "user-runtime-dir@.service": {"name": "user-runtime-dir@.service", "state": "unknown", "status": "static", "source": "systemd"}, "user@.service": {"name": "user@.service", "state": "unknown", "status": "static", "source": "systemd"}}}, "invocation": {"module_args": {}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203883.30757: done with _execute_module (service_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'service_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203883.30791: _low_level_execute_command(): starting 13731 1727203883.30803: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203881.6643684-18118-59095570388317/ > /dev/null 2>&1 && sleep 0' 13731 1727203883.32065: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203883.32197: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203883.32267: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203883.32327: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203883.32401: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203883.32463: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203883.34781: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203883.34785: stdout chunk (state=3): >>><<< 13731 1727203883.34787: stderr chunk (state=3): >>><<< 13731 1727203883.34790: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203883.34792: handler run complete 13731 1727203883.35132: variable 'ansible_facts' from source: unknown 13731 1727203883.35474: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203883.36884: variable 'ansible_facts' from source: unknown 13731 1727203883.37280: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203883.37283: attempt loop complete, returning result 13731 1727203883.37680: _execute() done 13731 1727203883.37684: dumping result to json 13731 1727203883.37686: done dumping result, returning 13731 1727203883.37688: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which services are running [028d2410-947f-82dc-c122-000000000fe5] 13731 1727203883.37691: sending task result for task 028d2410-947f-82dc-c122-000000000fe5 13731 1727203883.39222: done sending task result for task 028d2410-947f-82dc-c122-000000000fe5 13731 1727203883.39225: WORKER PROCESS EXITING ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203883.39339: no more pending results, returning what we have 13731 1727203883.39342: results queue empty 13731 1727203883.39343: checking for any_errors_fatal 13731 1727203883.39346: done checking for any_errors_fatal 13731 1727203883.39347: checking for max_fail_percentage 13731 1727203883.39348: done checking for max_fail_percentage 13731 1727203883.39349: checking to see if all hosts have failed and the running result is not ok 13731 1727203883.39350: done checking to see if all hosts have failed 13731 1727203883.39350: getting the remaining hosts for this loop 13731 1727203883.39352: done getting the remaining hosts for this loop 13731 1727203883.39355: getting the next task for host managed-node3 13731 1727203883.39360: done getting next task for host managed-node3 13731 1727203883.39363: ^ task is: TASK: fedora.linux_system_roles.network : Check which packages are installed 13731 1727203883.39370: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203883.39382: getting variables 13731 1727203883.39383: in VariableManager get_vars() 13731 1727203883.39414: Calling all_inventory to load vars for managed-node3 13731 1727203883.39417: Calling groups_inventory to load vars for managed-node3 13731 1727203883.39419: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203883.39429: Calling all_plugins_play to load vars for managed-node3 13731 1727203883.39432: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203883.39435: Calling groups_plugins_play to load vars for managed-node3 13731 1727203883.41806: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203883.43662: done with get_vars() 13731 1727203883.43692: done getting variables TASK [fedora.linux_system_roles.network : Check which packages are installed] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 Tuesday 24 September 2024 14:51:23 -0400 (0:00:01.815) 0:01:01.668 ***** 13731 1727203883.43800: entering _queue_task() for managed-node3/package_facts 13731 1727203883.44160: worker is 1 (out of 1 available) 13731 1727203883.44174: exiting _queue_task() for managed-node3/package_facts 13731 1727203883.44390: done queuing things up, now waiting for results queue to drain 13731 1727203883.44392: waiting for pending results... 13731 1727203883.44498: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which packages are installed 13731 1727203883.44695: in run() - task 028d2410-947f-82dc-c122-000000000fe6 13731 1727203883.44717: variable 'ansible_search_path' from source: unknown 13731 1727203883.44732: variable 'ansible_search_path' from source: unknown 13731 1727203883.44772: calling self._execute() 13731 1727203883.44871: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203883.44947: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203883.44950: variable 'omit' from source: magic vars 13731 1727203883.45269: variable 'ansible_distribution_major_version' from source: facts 13731 1727203883.45290: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203883.45301: variable 'omit' from source: magic vars 13731 1727203883.45391: variable 'omit' from source: magic vars 13731 1727203883.45429: variable 'omit' from source: magic vars 13731 1727203883.45477: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203883.45525: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203883.45548: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203883.45569: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203883.45587: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203883.45681: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203883.45685: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203883.45687: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203883.45753: Set connection var ansible_pipelining to False 13731 1727203883.45764: Set connection var ansible_shell_type to sh 13731 1727203883.45777: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203883.45789: Set connection var ansible_connection to ssh 13731 1727203883.45799: Set connection var ansible_shell_executable to /bin/sh 13731 1727203883.45809: Set connection var ansible_timeout to 10 13731 1727203883.45843: variable 'ansible_shell_executable' from source: unknown 13731 1727203883.45852: variable 'ansible_connection' from source: unknown 13731 1727203883.45929: variable 'ansible_module_compression' from source: unknown 13731 1727203883.45932: variable 'ansible_shell_type' from source: unknown 13731 1727203883.45934: variable 'ansible_shell_executable' from source: unknown 13731 1727203883.45937: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203883.45939: variable 'ansible_pipelining' from source: unknown 13731 1727203883.45941: variable 'ansible_timeout' from source: unknown 13731 1727203883.45943: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203883.46101: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203883.46118: variable 'omit' from source: magic vars 13731 1727203883.46128: starting attempt loop 13731 1727203883.46135: running the handler 13731 1727203883.46162: _low_level_execute_command(): starting 13731 1727203883.46177: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203883.46925: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203883.46994: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203883.47050: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203883.47085: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203883.47136: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203883.47247: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203883.48872: stdout chunk (state=3): >>>/root <<< 13731 1727203883.48954: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203883.48985: stdout chunk (state=3): >>><<< 13731 1727203883.49185: stderr chunk (state=3): >>><<< 13731 1727203883.49190: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203883.49194: _low_level_execute_command(): starting 13731 1727203883.49198: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301 `" && echo ansible-tmp-1727203883.4910989-18164-6394255790301="` echo /root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301 `" ) && sleep 0' 13731 1727203883.50336: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203883.50492: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203883.50619: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203883.50631: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203883.50655: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203883.50793: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203883.52579: stdout chunk (state=3): >>>ansible-tmp-1727203883.4910989-18164-6394255790301=/root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301 <<< 13731 1727203883.52881: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203883.52886: stdout chunk (state=3): >>><<< 13731 1727203883.52889: stderr chunk (state=3): >>><<< 13731 1727203883.52893: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203883.4910989-18164-6394255790301=/root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203883.52896: variable 'ansible_module_compression' from source: unknown 13731 1727203883.52899: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.package_facts-ZIP_DEFLATED 13731 1727203883.53281: variable 'ansible_facts' from source: unknown 13731 1727203883.53527: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/AnsiballZ_package_facts.py 13731 1727203883.53796: Sending initial data 13731 1727203883.53805: Sent initial data (160 bytes) 13731 1727203883.55072: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203883.55173: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203883.55220: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203883.55245: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203883.56762: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203883.56838: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/AnsiballZ_package_facts.py" <<< 13731 1727203883.56842: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpu51ux3ca /root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/AnsiballZ_package_facts.py <<< 13731 1727203883.56941: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpu51ux3ca" to remote "/root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/AnsiballZ_package_facts.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/AnsiballZ_package_facts.py" <<< 13731 1727203883.60363: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203883.60453: stderr chunk (state=3): >>><<< 13731 1727203883.60462: stdout chunk (state=3): >>><<< 13731 1727203883.60488: done transferring module to remote 13731 1727203883.60503: _low_level_execute_command(): starting 13731 1727203883.60533: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/ /root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/AnsiballZ_package_facts.py && sleep 0' 13731 1727203883.61616: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203883.61649: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203883.61667: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203883.61691: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203883.61733: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203883.61806: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203883.61822: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203883.61849: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203883.61967: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203883.63813: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203883.63826: stdout chunk (state=3): >>><<< 13731 1727203883.63828: stderr chunk (state=3): >>><<< 13731 1727203883.63933: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203883.63937: _low_level_execute_command(): starting 13731 1727203883.63939: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/AnsiballZ_package_facts.py && sleep 0' 13731 1727203883.65344: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203883.65358: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203883.65428: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203883.65522: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203883.65555: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203884.10429: stdout chunk (state=3): >>> {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "hunspell-filesystem": [{"name": "hunspell-filesystem", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "google-noto-fonts-common": [{"name": "google-noto-fonts-common", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-mono-vf-fonts": [{"name": "google-noto-sans-mono-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-vf-fonts": [{"name": "google-noto-sans-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-serif-vf-fonts": [{"name": "google-noto-serif-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-mono-vf-fonts": [{"name": "redhat-mono-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-text-vf-fonts": [{"name": "redhat-text-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "default-fonts-core-sans": [{"name": "default-fonts-core-sans", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-fonts-en": [{"name": "langpacks-fonts-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-ucode-firmware": [{"name": "amd-ucode-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "atheros-firmware": [{"name": "atheros-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "brcmfmac-firmware": [{"name": "brcmfmac-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "cirrus-audio-firmware": [{"name": "cirrus-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-audio-firmware": [{"name": "intel-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "mt7xxx-firmware": [{"name": "mt7xxx-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nxpwireless-firmware": [{"name": "nxpwireless-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "realtek-firmware": [{"name": "realtek-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tiwilink-firmware": [{"name": "tiwilink-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-gpu-firmware": [{"name": "amd-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-gpu-firmware": [{"name": "intel-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nvidia-gpu-firmware": [{"name": "nvidia-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "xkeyboard-config": [{"name": "xkeyboard-config", "version": "2.41", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-data": [{"name": "vim-data", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20240107", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.379", "release": "10.1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.14.5", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.18", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "21.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.2.26", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-ng-compat": [{"name": "zlib-ng-compat", "version": "2.1.6", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.19", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.5", "rele<<< 13731 1727203884.10749: stdout chunk (state=3): >>>ase": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.5", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.36", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.30", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.50", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.11", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.46.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.4", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "1.1", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "51.20230828cvs.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.6.2", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.1", "release": "9.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.17", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.9", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.10.0", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.5.0", "release": "1.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "hunspell-en-GB": [{"name": "hunspell-en-GB", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell-en-US": [{"name": "hunspell-en-US", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell": [{"name": "hunspell", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.6.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam-libs": [{"name": "pam-libs", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.69", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.15.0", "release": "3.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.19.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.80.4", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-libs": [{"name": "polkit-libs", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "101.2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.13", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.11.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.6.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.7", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.10", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.3", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.6", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.4", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.7.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.16", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.6.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.5", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "42.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.7", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sequoia": [{"name": "rpm-sequoia", "version": "1.6.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.29", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.79.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsecret": [{"name": "libsecret", "version": "0.21.2", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pinentry": [{"name": "pinentry", "version": "1.3.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libusb1": [{"name": "libusb1", "version": "1.0.27", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "4.0.4", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hunspell-en": [{"name": "hunspell-en", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.2.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm": [{"name": "gdbm", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libxkbcommon": [{"name": "libxkbcommon", "version": "1.7.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.23.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.7", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.10", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.8", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "35", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "661", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.6", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "memstrack": [{"name": "memstrack", "version": "0.2.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.25.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.15", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "duktape": [{"name": "duktape", "version": "2.7.0", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "22.el10.gating_test1", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse3-libs": [{"name": "fuse3-libs", "version": "3.16.2", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-envsubst": [{"name": "gettext-envsubst", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-runtime": [{"name": "gettext-runtime", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "58", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.1.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.11.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.14.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "28.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.62.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.7", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-gssapi": [{"name": "cyrus-sasl-gssapi", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11^20190603git9e74f2d", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "polkit": [{"name": "polkit", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-pkla-compat": [{"name": "polkit-pkla-compat", "version": "0.1", "release": "29.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "audit-rules": [{"name": "audit-rules", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20240905", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hasher": [{"name": "libkcapi-hasher", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.22.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "makedumpfile": [{"name": "makedumpfile", "version": "1.7.5", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.9.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.18.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss-fapi": [{"name": "tpm2-tss-fapi", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tools": [{"name": "tpm2-tools", "version": "5.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "76.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.81", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.15.0", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.32", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-krb5-common": [{"name": "sssd-krb5-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpdecimal": [{"name": "mpdecimal", "version": "2.5.1", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.3.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo-python-plugin": [{"name": "sudo-python-plugin", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.16.0", "release": "15.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.2", "release": "14.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "235", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.9", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.7.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.29", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kdump-utils": [{"name": "kdump-utils", "version": "1.0.43", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.13.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-libs": [{"name": "pcsc-lite-libs", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-ccid": [<<< 13731 1727203884.10823: stdout chunk (state=3): >>>{"name": "pcsc-lite-ccid", "version": "1.6.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite": [{"name": "pcsc-lite", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2-smime": [{"name": "gnupg2-smime", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.3", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.24", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.14.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "1.2304.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2408.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "8.el10", "epoch": 18, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.12.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.2.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.20", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwlwifi-dvm-firmware": [{"name": "iwlwifi-dvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwlwifi-mvm-firmware": [{"name": "iwlwifi-mvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "37.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.5", "release": "0.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.20", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.59", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.89", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.189", "release": "511.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.15", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.27", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2024.001", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20231213", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.61.000", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-locale": [{"name": "perl-locale", "version": "1.12", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.42", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.350", "release": "510.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.085", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.68", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "2.20", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.22", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.45", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.088", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.038", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.09", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "5.01", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.18", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.31", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.29", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.55", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.03", "release": "510.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.38", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.86", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.63", "release": "510.el10", "epoch": 5, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.32", "release": "510.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.37", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.241", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.58", "release": "2.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.54", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.78", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.91", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.56", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.17", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.21", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.44", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "23.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xxd": [{"name": "xxd", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.39", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "5.2.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "24.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.35", "release": "4.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.10", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.24.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.4.1", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.9.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.3.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpds-py": [{"name": "python3-rpds-py", "version": "0.17.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "23.2.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-referencing": [{"name": "python3-referencing", "version": "0.31.1", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.19", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema-specifications": [{"name": "python3-jsonschema-specifications", "version": "2023.11.2", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "4.19.1", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.5", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.2.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "2.1.3", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "3.1.4", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.3", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.33", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.9.0", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.8", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "69.0.3", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.5.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "6.0.1", "release": "18.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-charset-normalizer": [{"name": "python3-charset-normalizer", "version": "3.3.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.32.3", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dhcpcd": [{"name": "dhcpcd", "version": "10.0.6", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "24.1.4", "release": "17.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "device-mapper-event-libs": [{"name": "device-mapper-event-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "libaio": [{"name": "libaio", "version": "0.3.111", "release": "20.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-event": [{"name": "device-mapper-event", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "lvm2-libs": [{"name": "lvm2-libs", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper-persistent-data": [{"name": "device-mapper-persistent-data", "version": "1.0.11", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lvm2": [{"name": "lvm2", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.17", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.90", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} <<< 13731 1727203884.12460: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203884.12495: stderr chunk (state=3): >>><<< 13731 1727203884.12518: stdout chunk (state=3): >>><<< 13731 1727203884.12580: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"packages": {"libgcc": [{"name": "libgcc", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "linux-firmware-whence": [{"name": "linux-firmware-whence", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tzdata": [{"name": "tzdata", "version": "2024a", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "fonts-filesystem": [{"name": "fonts-filesystem", "version": "2.0.5", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "hunspell-filesystem": [{"name": "hunspell-filesystem", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "google-noto-fonts-common": [{"name": "google-noto-fonts-common", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-mono-vf-fonts": [{"name": "google-noto-sans-mono-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-sans-vf-fonts": [{"name": "google-noto-sans-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "google-noto-serif-vf-fonts": [{"name": "google-noto-serif-vf-fonts", "version": "20240401", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-mono-vf-fonts": [{"name": "redhat-mono-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "redhat-text-vf-fonts": [{"name": "redhat-text-vf-fonts", "version": "4.0.3", "release": "12.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "default-fonts-core-sans": [{"name": "default-fonts-core-sans", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-fonts-en": [{"name": "langpacks-fonts-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-ucode-firmware": [{"name": "amd-ucode-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "atheros-firmware": [{"name": "atheros-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "brcmfmac-firmware": [{"name": "brcmfmac-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "cirrus-audio-firmware": [{"name": "cirrus-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-audio-firmware": [{"name": "intel-audio-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "mt7xxx-firmware": [{"name": "mt7xxx-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nxpwireless-firmware": [{"name": "nxpwireless-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "realtek-firmware": [{"name": "realtek-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tiwilink-firmware": [{"name": "tiwilink-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "amd-gpu-firmware": [{"name": "amd-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "intel-gpu-firmware": [{"name": "intel-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "nvidia-gpu-firmware": [{"name": "nvidia-gpu-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "linux-firmware": [{"name": "linux-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "xkeyboard-config": [{"name": "xkeyboard-config", "version": "2.41", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "gawk-all-langpacks": [{"name": "gawk-all-langpacks", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-data": [{"name": "vim-data", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "publicsuffix-list-dafsa": [{"name": "publicsuffix-list-dafsa", "version": "20240107", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "pcre2-syntax": [{"name": "pcre2-syntax", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "noarch", "source": "rpm"}], "ncurses-base": [{"name": "ncurses-base", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libssh-config": [{"name": "libssh-config", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-misc": [{"name": "kbd-misc", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kbd-legacy": [{"name": "kbd-legacy", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hwdata": [{"name": "hwdata", "version": "0.379", "release": "10.1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "firewalld-filesystem": [{"name": "firewalld-filesystem", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf-data": [{"name": "dnf-data", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "coreutils-common": [{"name": "coreutils-common", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "centos-gpg-keys": [{"name": "centos-gpg-keys", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-repos": [{"name": "centos-stream-repos", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "centos-stream-release": [{"name": "centos-stream-release", "version": "10.0", "release": "0.19.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "setup": [{"name": "setup", "version": "2.14.5", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "filesystem": [{"name": "filesystem", "version": "3.18", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "basesystem": [{"name": "basesystem", "version": "11", "release": "21.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "glibc-gconv-extra": [{"name": "glibc-gconv-extra", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-langpack-en": [{"name": "glibc-langpack-en", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc-common": [{"name": "glibc-common", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glibc": [{"name": "glibc", "version": "2.39", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses-libs": [{"name": "ncurses-libs", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bash": [{"name": "bash", "version": "5.2.26", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "zlib-ng-compat": [{"name": "zlib-ng-compat", "version": "2.1.6", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libuuid": [{"name": "libuuid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz-libs": [{"name": "xz-libs", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libblkid": [{"name": "libblkid", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libstdc++": [{"name": "libstdc++", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "popt": [{"name": "popt", "version": "1.19", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libzstd": [{"name": "libzstd", "version": "1.5.5", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libelf": [{"name": "elfutils-libelf", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "readline": [{"name": "readline", "version": "8.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "bzip2-libs": [{"name": "bzip2-libs", "version": "1.0.8", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcom_err": [{"name": "libcom_err", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmnl": [{"name": "libmnl", "version": "1.0.5", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxcrypt": [{"name": "libxcrypt", "version": "4.4.36", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crypto-policies": [{"name": "crypto-policies", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "alternatives": [{"name": "alternatives", "version": "1.30", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libxml2": [{"name": "libxml2", "version": "2.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng": [{"name": "libcap-ng", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit-libs": [{"name": "audit-libs", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgpg-error": [{"name": "libgpg-error", "version": "1.50", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtalloc": [{"name": "libtalloc", "version": "2.4.2", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcre2": [{"name": "pcre2", "version": "10.44", "release": "1.el10.2", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grep": [{"name": "grep", "version": "3.11", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sqlite-libs": [{"name": "sqlite-libs", "version": "3.46.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm-libs": [{"name": "gdbm-libs", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libffi": [{"name": "libffi", "version": "3.4.4", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libunistring": [{"name": "libunistring", "version": "1.1", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libidn2": [{"name": "libidn2", "version": "2.3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-common": [{"name": "grub2-common", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libedit": [{"name": "libedit", "version": "3.1", "release": "51.20230828cvs.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "expat": [{"name": "expat", "version": "2.6.2", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gmp": [{"name": "gmp", "version": "6.2.1", "release": "9.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "jansson": [{"name": "jansson", "version": "2.14", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "json-c": [{"name": "json-c", "version": "0.17", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libattr": [{"name": "libattr", "version": "2.5.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libacl": [{"name": "libacl", "version": "2.3.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsepol": [{"name": "libsepol", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libselinux": [{"name": "libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sed": [{"name": "sed", "version": "4.9", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmount": [{"name": "libmount", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsmartcols": [{"name": "libsmartcols", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "findutils": [{"name": "findutils", "version": "4.10.0", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libsemanage": [{"name": "libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtevent": [{"name": "libtevent", "version": "0.16.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libassuan": [{"name": "libassuan", "version": "2.5.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbpf": [{"name": "libbpf", "version": "1.5.0", "release": "1.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "hunspell-en-GB": [{"name": "hunspell-en-GB", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell-en-US": [{"name": "hunspell-en-US", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "hunspell": [{"name": "hunspell", "version": "1.7.2", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfdisk": [{"name": "libfdisk", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils-libs": [{"name": "keyutils-libs", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libeconf": [{"name": "libeconf", "version": "0.6.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pam-libs": [{"name": "pam-libs", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap": [{"name": "libcap", "version": "2.69", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-libs": [{"name": "systemd-libs", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "shadow-utils": [{"name": "shadow-utils", "version": "4.15.0", "release": "3.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "util-linux-core": [{"name": "util-linux-core", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-libs": [{"name": "dbus-libs", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libtasn1": [{"name": "libtasn1", "version": "4.19.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit": [{"name": "p11-kit", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "p11-kit-trust": [{"name": "p11-kit-trust", "version": "0.25.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnutls": [{"name": "gnutls", "version": "3.8.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "glib2": [{"name": "glib2", "version": "2.80.4", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-libs": [{"name": "polkit-libs", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-libnm": [{"name": "NetworkManager-libnm", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "openssl-libs": [{"name": "openssl-libs", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "coreutils": [{"name": "coreutils", "version": "9.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ca-certificates": [{"name": "ca-certificates", "version": "2024.2.69_v8.0.303", "release": "101.2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "tpm2-tss": [{"name": "tpm2-tss", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gzip": [{"name": "gzip", "version": "1.13", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod": [{"name": "kmod", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kmod-libs": [{"name": "kmod-libs", "version": "31", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib": [{"name": "cracklib", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-lib": [{"name": "cyrus-sasl-lib", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgcrypt": [{"name": "libgcrypt", "version": "1.11.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libksba": [{"name": "libksba", "version": "1.6.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnftnl": [{"name": "libnftnl", "version": "1.2.7", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file-libs": [{"name": "file-libs", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "file": [{"name": "file", "version": "5.45", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "diffutils": [{"name": "diffutils", "version": "3.10", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbasicobjects": [{"name": "libbasicobjects", "version": "0.1.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcollection": [{"name": "libcollection", "version": "0.7.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdhash": [{"name": "libdhash", "version": "0.5.0", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnl3": [{"name": "libnl3", "version": "3.9.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libref_array": [{"name": "libref_array", "version": "0.1.5", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libseccomp": [{"name": "libseccomp", "version": "2.5.3", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_idmap": [{"name": "libsss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libtdb": [{"name": "libtdb", "version": "1.4.10", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lua-libs": [{"name": "lua-libs", "version": "5.4.6", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lz4-libs": [{"name": "lz4-libs", "version": "1.9.4", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libarchive": [{"name": "libarchive", "version": "3.7.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lzo": [{"name": "lzo", "version": "2.10", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "npth": [{"name": "npth", "version": "1.6", "release": "19.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "numactl-libs": [{"name": "numactl-libs", "version": "2.0.16", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "squashfs-tools": [{"name": "squashfs-tools", "version": "4.6.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cracklib-dicts": [{"name": "cracklib-dicts", "version": "2.9.11", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpwquality": [{"name": "libpwquality", "version": "1.4.5", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ima-evm-utils": [{"name": "ima-evm-utils", "version": "1.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip-wheel": [{"name": "python3-pip-wheel", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "which": [{"name": "which", "version": "2.21", "release": "42.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libevent": [{"name": "libevent", "version": "2.1.12", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openldap": [{"name": "openldap", "version": "2.6.7", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_certmap": [{"name": "libsss_certmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sequoia": [{"name": "rpm-sequoia", "version": "1.6.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-audit": [{"name": "rpm-plugin-audit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-libs": [{"name": "rpm-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsolv": [{"name": "libsolv", "version": "0.7.29", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-systemd-inhibit": [{"name": "rpm-plugin-systemd-inhibit", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gobject-introspection": [{"name": "gobject-introspection", "version": "1.79.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsecret": [{"name": "libsecret", "version": "0.21.2", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pinentry": [{"name": "pinentry", "version": "1.3.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libusb1": [{"name": "libusb1", "version": "1.0.27", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "procps-ng": [{"name": "procps-ng", "version": "4.0.4", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kbd": [{"name": "kbd", "version": "2.6.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "hunspell-en": [{"name": "hunspell-en", "version": "0.20201207", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libselinux-utils": [{"name": "libselinux-utils", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-libs": [{"name": "gettext-libs", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpfr": [{"name": "mpfr", "version": "4.2.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gawk": [{"name": "gawk", "version": "5.3.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcomps": [{"name": "libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc-modules": [{"name": "grub2-pc-modules", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "libpsl": [{"name": "libpsl", "version": "0.21.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gdbm": [{"name": "gdbm", "version": "1.23", "release": "8.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "pam": [{"name": "pam", "version": "1.6.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xz": [{"name": "xz", "version": "5.6.2", "release": "2.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libxkbcommon": [{"name": "libxkbcommon", "version": "1.7.0", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "groff-base": [{"name": "groff-base", "version": "1.23.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ethtool": [{"name": "ethtool", "version": "6.7", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "ipset-libs": [{"name": "ipset-libs", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ipset": [{"name": "ipset", "version": "7.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs-libs": [{"name": "e2fsprogs-libs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libss": [{"name": "libss", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "snappy": [{"name": "snappy", "version": "1.1.10", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pigz": [{"name": "pigz", "version": "2.8", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus-common": [{"name": "dbus-common", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "dbus-broker": [{"name": "dbus-broker", "version": "35", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dbus": [{"name": "dbus", "version": "1.14.10", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "hostname": [{"name": "hostname", "version": "3.23", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-tools-libs": [{"name": "kernel-tools-libs", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "less": [{"name": "less", "version": "661", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "psmisc": [{"name": "psmisc", "version": "23.6", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute": [{"name": "iproute", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "memstrack": [{"name": "memstrack", "version": "0.2.5", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "c-ares": [{"name": "c-ares", "version": "1.25.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cpio": [{"name": "cpio", "version": "2.15", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "duktape": [{"name": "duktape", "version": "2.7.0", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse-libs": [{"name": "fuse-libs", "version": "2.9.9", "release": "22.el10.gating_test1", "epoch": null, "arch": "x86_64", "source": "rpm"}], "fuse3-libs": [{"name": "fuse3-libs", "version": "3.16.2", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-envsubst": [{"name": "gettext-envsubst", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gettext-runtime": [{"name": "gettext-runtime", "version": "0.22.5", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "inih": [{"name": "inih", "version": "58", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libbrotli": [{"name": "libbrotli", "version": "1.1.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcbor": [{"name": "libcbor", "version": "0.11.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfido2": [{"name": "libfido2", "version": "1.14.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libgomp": [{"name": "libgomp", "version": "14.2.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libndp": [{"name": "libndp", "version": "1.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfnetlink": [{"name": "libnfnetlink", "version": "1.0.1", "release": "28.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnetfilter_conntrack": [{"name": "libnetfilter_conntrack", "version": "1.0.9", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-libs": [{"name": "iptables-libs", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iptables-nft": [{"name": "iptables-nft", "version": "1.8.10", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nftables": [{"name": "nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libnghttp2": [{"name": "libnghttp2", "version": "1.62.1", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpath_utils": [{"name": "libpath_utils", "version": "0.2.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libini_config": [{"name": "libini_config", "version": "1.3.1", "release": "57.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libpipeline": [{"name": "libpipeline", "version": "1.5.7", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_nss_idmap": [{"name": "libsss_nss_idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsss_sudo": [{"name": "libsss_sudo", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "liburing": [{"name": "liburing", "version": "2.5", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto": [{"name": "libverto", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "krb5-libs": [{"name": "krb5-libs", "version": "1.21.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cyrus-sasl-gssapi": [{"name": "cyrus-sasl-gssapi", "version": "2.1.28", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libssh": [{"name": "libssh", "version": "0.10.6", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcurl": [{"name": "libcurl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect-libs": [{"name": "authselect-libs", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cryptsetup-libs": [{"name": "cryptsetup-libs", "version": "2.7.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-libs": [{"name": "device-mapper-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper": [{"name": "device-mapper", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "elfutils-debuginfod-client": [{"name": "elfutils-debuginfod-client", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-libs": [{"name": "elfutils-libs", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "elfutils-default-yama-scope": [{"name": "elfutils-default-yama-scope", "version": "0.191", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libutempter": [{"name": "libutempter", "version": "1.2.1", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-pam": [{"name": "systemd-pam", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "util-linux": [{"name": "util-linux", "version": "2.40.2", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd": [{"name": "systemd", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools-minimal": [{"name": "grub2-tools-minimal", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "cronie-anacron": [{"name": "cronie-anacron", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cronie": [{"name": "cronie", "version": "1.7.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "crontabs": [{"name": "crontabs", "version": "1.11^20190603git9e74f2d", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "polkit": [{"name": "polkit", "version": "125", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "polkit-pkla-compat": [{"name": "polkit-pkla-compat", "version": "0.1", "release": "29.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh": [{"name": "openssh", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils-gold": [{"name": "binutils-gold", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "binutils": [{"name": "binutils", "version": "2.41", "release": "48.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-service": [{"name": "initscripts-service", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "audit-rules": [{"name": "audit-rules", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "audit": [{"name": "audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iputils": [{"name": "iputils", "version": "20240905", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi": [{"name": "libkcapi", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hasher": [{"name": "libkcapi-hasher", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libkcapi-hmaccalc": [{"name": "libkcapi-hmaccalc", "version": "1.5.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "logrotate": [{"name": "logrotate", "version": "3.22.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "makedumpfile": [{"name": "makedumpfile", "version": "1.7.5", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-build-libs": [{"name": "rpm-build-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kpartx": [{"name": "kpartx", "version": "0.9.9", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "curl": [{"name": "curl", "version": "8.9.1", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm": [{"name": "rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "policycoreutils": [{"name": "policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "selinux-policy": [{"name": "selinux-policy", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "selinux-policy-targeted": [{"name": "selinux-policy-targeted", "version": "40.13.9", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "librepo": [{"name": "librepo", "version": "1.18.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tss-fapi": [{"name": "tpm2-tss-fapi", "version": "4.1.3", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tpm2-tools": [{"name": "tpm2-tools", "version": "5.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grubby": [{"name": "grubby", "version": "8.40", "release": "76.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "systemd-udev": [{"name": "systemd-udev", "version": "256", "release": "14.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut": [{"name": "dracut", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "os-prober": [{"name": "os-prober", "version": "1.81", "release": "8.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-tools": [{"name": "grub2-tools", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules-core": [{"name": "kernel-modules-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel-core": [{"name": "kernel-core", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager": [{"name": "NetworkManager", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "kernel-modules": [{"name": "kernel-modules", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-squash": [{"name": "dracut-squash", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-client": [{"name": "sssd-client", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libyaml": [{"name": "libyaml", "version": "0.2.5", "release": "15.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libmodulemd": [{"name": "libmodulemd", "version": "2.15.0", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libdnf": [{"name": "libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lmdb-libs": [{"name": "lmdb-libs", "version": "0.9.32", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libldb": [{"name": "libldb", "version": "2.9.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-common": [{"name": "sssd-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-krb5-common": [{"name": "sssd-krb5-common", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "mpdecimal": [{"name": "mpdecimal", "version": "2.5.1", "release": "11.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python-unversioned-command": [{"name": "python-unversioned-command", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3": [{"name": "python3", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libs": [{"name": "python3-libs", "version": "3.12.5", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dbus": [{"name": "python3-dbus", "version": "1.3.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libdnf": [{"name": "python3-libdnf", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-hawkey": [{"name": "python3-hawkey", "version": "0.73.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-gobject-base-noarch": [{"name": "python3-gobject-base-noarch", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-gobject-base": [{"name": "python3-gobject-base", "version": "3.46.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-libcomps": [{"name": "python3-libcomps", "version": "0.1.21", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo": [{"name": "sudo", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sudo-python-plugin": [{"name": "sudo-python-plugin", "version": "1.9.15", "release": "7.p5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-nftables": [{"name": "python3-nftables", "version": "1.0.9", "release": "4.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "python3-firewall": [{"name": "python3-firewall", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-six": [{"name": "python3-six", "version": "1.16.0", "release": "15.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dateutil": [{"name": "python3-dateutil", "version": "2.8.2", "release": "14.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "python3-systemd": [{"name": "python3-systemd", "version": "235", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libcap-ng-python3": [{"name": "libcap-ng-python3", "version": "0.8.4", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "oniguruma": [{"name": "oniguruma", "version": "6.9.9", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "jq": [{"name": "jq", "version": "1.7.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dracut-network": [{"name": "dracut-network", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kexec-tools": [{"name": "kexec-tools", "version": "2.0.29", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kdump-utils": [{"name": "kdump-utils", "version": "1.0.43", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pciutils-libs": [{"name": "pciutils-libs", "version": "3.13.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-libs": [{"name": "pcsc-lite-libs", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite-ccid": [{"name": "pcsc-lite-ccid", "version": "1.6.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "pcsc-lite": [{"name": "pcsc-lite", "version": "2.2.3", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2-smime": [{"name": "gnupg2-smime", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gnupg2": [{"name": "gnupg2", "version": "2.4.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-sign-libs": [{"name": "rpm-sign-libs", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpm": [{"name": "python3-rpm", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-dnf": [{"name": "python3-dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnf": [{"name": "dnf", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-dnf-plugins-core": [{"name": "python3-dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "sg3_utils-libs": [{"name": "sg3_utils-libs", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "slang": [{"name": "slang", "version": "2.3.3", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "newt": [{"name": "newt", "version": "0.52.24", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "userspace-rcu": [{"name": "userspace-rcu", "version": "0.14.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libestr": [{"name": "libestr", "version": "0.1.11", "release": "10.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libfastjson": [{"name": "libfastjson", "version": "1.2304.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "langpacks-core-en": [{"name": "langpacks-core-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "langpacks-en": [{"name": "langpacks-en", "version": "4.1", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rsyslog": [{"name": "rsyslog", "version": "8.2408.0", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xfsprogs": [{"name": "xfsprogs", "version": "6.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "NetworkManager-tui": [{"name": "NetworkManager-tui", "version": "1.48.10", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "sg3_utils": [{"name": "sg3_utils", "version": "1.48", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "dnf-plugins-core": [{"name": "dnf-plugins-core", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "yum": [{"name": "yum", "version": "4.20.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "kernel-tools": [{"name": "kernel-tools", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "firewalld": [{"name": "firewalld", "version": "2.2.1", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "crypto-policies-scripts": [{"name": "crypto-policies-scripts", "version": "20240822", "release": "1.git367040b.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libselinux": [{"name": "python3-libselinux", "version": "3.7", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "sssd-kcm": [{"name": "sssd-kcm", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "kernel": [{"name": "kernel", "version": "6.11.0", "release": "25.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "grub2-pc": [{"name": "grub2-pc", "version": "2.06", "release": "127.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dracut-config-rescue": [{"name": "dracut-config-rescue", "version": "102", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-clients": [{"name": "openssh-clients", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "openssh-server": [{"name": "openssh-server", "version": "9.8p1", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "chrony": [{"name": "chrony", "version": "4.6", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "microcode_ctl": [{"name": "microcode_ctl", "version": "20240531", "release": "1.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "qemu-guest-agent": [{"name": "qemu-guest-agent", "version": "9.0.0", "release": "8.el10", "epoch": 18, "arch": "x86_64", "source": "rpm"}], "parted": [{"name": "parted", "version": "3.6", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "authselect": [{"name": "authselect", "version": "1.5.0", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "man-db": [{"name": "man-db", "version": "2.12.0", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iproute-tc": [{"name": "iproute-tc", "version": "6.7.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "e2fsprogs": [{"name": "e2fsprogs", "version": "1.47.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "initscripts-rename-device": [{"name": "initscripts-rename-device", "version": "10.26", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rpm-plugin-selinux": [{"name": "rpm-plugin-selinux", "version": "4.19.1.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "irqbalance": [{"name": "irqbalance", "version": "1.9.4", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "prefixdevname": [{"name": "prefixdevname", "version": "0.2.0", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-minimal": [{"name": "vim-minimal", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "lshw": [{"name": "lshw", "version": "B.02.20", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "ncurses": [{"name": "ncurses", "version": "6.4", "release": "13.20240127.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libsysfs": [{"name": "libsysfs", "version": "2.1.1", "release": "13.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lsscsi": [{"name": "lsscsi", "version": "0.32", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "iwlwifi-dvm-firmware": [{"name": "iwlwifi-dvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "iwlwifi-mvm-firmware": [{"name": "iwlwifi-mvm-firmware", "version": "20240910", "release": "7.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rootfiles": [{"name": "rootfiles", "version": "8.1", "release": "37.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "libtirpc": [{"name": "libtirpc", "version": "1.3.5", "release": "0.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "git-core": [{"name": "git-core", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libnfsidmap": [{"name": "libnfsidmap", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "git-core-doc": [{"name": "git-core-doc", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "rpcbind": [{"name": "rpcbind", "version": "1.2.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Digest": [{"name": "perl-Digest", "version": "1.20", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Digest-MD5": [{"name": "perl-Digest-MD5", "version": "2.59", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-B": [{"name": "perl-B", "version": "1.89", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-FileHandle": [{"name": "perl-FileHandle", "version": "2.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Data-Dumper": [{"name": "perl-Data-Dumper", "version": "2.189", "release": "511.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-libnet": [{"name": "perl-libnet", "version": "3.15", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-URI": [{"name": "perl-URI", "version": "5.27", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-AutoLoader": [{"name": "perl-AutoLoader", "version": "5.74", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Text-Tabs+Wrap": [{"name": "perl-Text-Tabs+Wrap", "version": "2024.001", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Mozilla-CA": [{"name": "perl-Mozilla-CA", "version": "20231213", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-if": [{"name": "perl-if", "version": "0.61.000", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-locale": [{"name": "perl-locale", "version": "1.12", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-IP": [{"name": "perl-IO-Socket-IP", "version": "0.42", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Time-Local": [{"name": "perl-Time-Local", "version": "1.350", "release": "510.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "perl-File-Path": [{"name": "perl-File-Path", "version": "2.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Escapes": [{"name": "perl-Pod-Escapes", "version": "1.07", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-IO-Socket-SSL": [{"name": "perl-IO-Socket-SSL", "version": "2.085", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Net-SSLeay": [{"name": "perl-Net-SSLeay", "version": "1.94", "release": "6.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Class-Struct": [{"name": "perl-Class-Struct", "version": "0.68", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Term-ANSIColor": [{"name": "perl-Term-ANSIColor", "version": "5.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-POSIX": [{"name": "perl-POSIX", "version": "2.20", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IPC-Open3": [{"name": "perl-IPC-Open3", "version": "1.22", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-Temp": [{"name": "perl-File-Temp", "version": "0.231.100", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Term-Cap": [{"name": "perl-Term-Cap", "version": "1.18", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Pod-Simple": [{"name": "perl-Pod-Simple", "version": "3.45", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-HTTP-Tiny": [{"name": "perl-HTTP-Tiny", "version": "0.088", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Socket": [{"name": "perl-Socket", "version": "2.038", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-SelectSaver": [{"name": "perl-SelectSaver", "version": "1.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Symbol": [{"name": "perl-Symbol", "version": "1.09", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-File-stat": [{"name": "perl-File-stat", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-podlators": [{"name": "perl-podlators", "version": "5.01", "release": "510.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Pod-Perldoc": [{"name": "perl-Pod-Perldoc", "version": "3.28.01", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Fcntl": [{"name": "perl-Fcntl", "version": "1.18", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Text-ParseWords": [{"name": "perl-Text-ParseWords", "version": "3.31", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-base": [{"name": "perl-base", "version": "2.27", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-mro": [{"name": "perl-mro", "version": "1.29", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-IO": [{"name": "perl-IO", "version": "1.55", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-overloading": [{"name": "perl-overloading", "version": "0.02", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Pod-Usage": [{"name": "perl-Pod-Usage", "version": "2.03", "release": "510.el10", "epoch": 4, "arch": "noarch", "source": "rpm"}], "perl-Errno": [{"name": "perl-Errno", "version": "1.38", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-File-Basename": [{"name": "perl-File-Basename", "version": "2.86", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Std": [{"name": "perl-Getopt-Std", "version": "1.14", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-MIME-Base64": [{"name": "perl-MIME-Base64", "version": "3.16", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-Scalar-List-Utils": [{"name": "perl-Scalar-List-Utils", "version": "1.63", "release": "510.el10", "epoch": 5, "arch": "x86_64", "source": "rpm"}], "perl-constant": [{"name": "perl-constant", "version": "1.33", "release": "511.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Storable": [{"name": "perl-Storable", "version": "3.32", "release": "510.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "perl-overload": [{"name": "perl-overload", "version": "1.37", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-parent": [{"name": "perl-parent", "version": "0.241", "release": "511.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-vars": [{"name": "perl-vars", "version": "1.05", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-Getopt-Long": [{"name": "perl-Getopt-Long", "version": "2.58", "release": "2.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-Carp": [{"name": "perl-Carp", "version": "1.54", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-Exporter": [{"name": "perl-Exporter", "version": "5.78", "release": "510.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "perl-PathTools": [{"name": "perl-PathTools", "version": "3.91", "release": "510.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-DynaLoader": [{"name": "perl-DynaLoader", "version": "1.56", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-NDBM_File": [{"name": "perl-NDBM_File", "version": "1.17", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Encode": [{"name": "perl-Encode", "version": "3.21", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-libs": [{"name": "perl-libs", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-interpreter": [{"name": "perl-interpreter", "version": "5.40.0", "release": "510.el10", "epoch": 4, "arch": "x86_64", "source": "rpm"}], "perl-Error": [{"name": "perl-Error", "version": "0.17029", "release": "17.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "perl-File-Find": [{"name": "perl-File-Find", "version": "1.44", "release": "510.el10", "epoch": 0, "arch": "noarch", "source": "rpm"}], "perl-TermReadKey": [{"name": "perl-TermReadKey", "version": "2.38", "release": "23.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "perl-lib": [{"name": "perl-lib", "version": "0.65", "release": "510.el10", "epoch": 0, "arch": "x86_64", "source": "rpm"}], "perl-Git": [{"name": "perl-Git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "git": [{"name": "git", "version": "2.45.2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "xxd": [{"name": "xxd", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "libxslt": [{"name": "libxslt", "version": "1.1.39", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-lxml": [{"name": "python3-lxml", "version": "5.2.1", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "yum-utils": [{"name": "yum-utils", "version": "4.7.0", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "vim-filesystem": [{"name": "vim-filesystem", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "noarch", "source": "rpm"}], "vim-common": [{"name": "vim-common", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "time": [{"name": "time", "version": "1.9", "release": "24.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "tar": [{"name": "tar", "version": "1.35", "release": "4.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "quota-nls": [{"name": "quota-nls", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "noarch", "source": "rpm"}], "quota": [{"name": "quota", "version": "4.09", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "nettle": [{"name": "nettle", "version": "3.10", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "wget": [{"name": "wget", "version": "1.24.5", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "make": [{"name": "make", "version": "4.4.1", "release": "7.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "libev": [{"name": "libev", "version": "4.33", "release": "12.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "libverto-libev": [{"name": "libverto-libev", "version": "0.3.2", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "gssproxy": [{"name": "gssproxy", "version": "0.9.2", "release": "7.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "keyutils": [{"name": "keyutils", "version": "1.6.3", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "nfs-utils": [{"name": "nfs-utils", "version": "2.7.1", "release": "1.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "bc": [{"name": "bc", "version": "1.07.1", "release": "22.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "beakerlib-redhat": [{"name": "beakerlib-redhat", "version": "1", "release": "35.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "beakerlib": [{"name": "beakerlib", "version": "1.29.3", "release": "1.el9", "epoch": null, "arch": "noarch", "source": "rpm"}], "restraint": [{"name": "restraint", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "restraint-rhts": [{"name": "restraint-rhts", "version": "0.4.4", "release": "1.el9", "epoch": null, "arch": "x86_64", "source": "rpm"}], "vim-enhanced": [{"name": "vim-enhanced", "version": "9.1.083", "release": "2.el10", "epoch": 2, "arch": "x86_64", "source": "rpm"}], "sssd-nfs-idmap": [{"name": "sssd-nfs-idmap", "version": "2.10.0~beta2", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rsync": [{"name": "rsync", "version": "3.3.0", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-rpds-py": [{"name": "python3-rpds-py", "version": "0.17.1", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-attrs": [{"name": "python3-attrs", "version": "23.2.0", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-referencing": [{"name": "python3-referencing", "version": "0.31.1", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-idna": [{"name": "python3-idna", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-urllib3": [{"name": "python3-urllib3", "version": "1.26.19", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema-specifications": [{"name": "python3-jsonschema-specifications", "version": "2023.11.2", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonschema": [{"name": "python3-jsonschema", "version": "4.19.1", "release": "6.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyserial": [{"name": "python3-pyserial", "version": "3.5", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-oauthlib": [{"name": "python3-oauthlib", "version": "3.2.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-markupsafe": [{"name": "python3-markupsafe", "version": "2.1.3", "release": "5.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jinja2": [{"name": "python3-jinja2", "version": "3.1.4", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-libsemanage": [{"name": "python3-libsemanage", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-jsonpointer": [{"name": "python3-jsonpointer", "version": "2.3", "release": "8.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-jsonpatch": [{"name": "python3-jsonpatch", "version": "1.33", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-distro": [{"name": "python3-distro", "version": "1.9.0", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-configobj": [{"name": "python3-configobj", "version": "5.0.8", "release": "9.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-audit": [{"name": "python3-audit", "version": "4.0", "release": "9.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "checkpolicy": [{"name": "checkpolicy", "version": "3.7", "release": "1.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-setuptools": [{"name": "python3-setuptools", "version": "69.0.3", "release": "5.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-setools": [{"name": "python3-setools", "version": "4.5.1", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-policycoreutils": [{"name": "python3-policycoreutils", "version": "3.7", "release": "2.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-pyyaml": [{"name": "python3-pyyaml", "version": "6.0.1", "release": "18.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-charset-normalizer": [{"name": "python3-charset-normalizer", "version": "3.3.2", "release": "4.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "python3-requests": [{"name": "python3-requests", "version": "2.32.3", "release": "1.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "openssl": [{"name": "openssl", "version": "3.2.2", "release": "12.el10", "epoch": 1, "arch": "x86_64", "source": "rpm"}], "dhcpcd": [{"name": "dhcpcd", "version": "10.0.6", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "cloud-init": [{"name": "cloud-init", "version": "24.1.4", "release": "17.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "device-mapper-event-libs": [{"name": "device-mapper-event-libs", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "libaio": [{"name": "libaio", "version": "0.3.111", "release": "20.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "device-mapper-event": [{"name": "device-mapper-event", "version": "1.02.198", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "lvm2-libs": [{"name": "lvm2-libs", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "device-mapper-persistent-data": [{"name": "device-mapper-persistent-data", "version": "1.0.11", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "lvm2": [{"name": "lvm2", "version": "2.03.24", "release": "2.el10", "epoch": 10, "arch": "x86_64", "source": "rpm"}], "cloud-utils-growpart": [{"name": "cloud-utils-growpart", "version": "0.33", "release": "10.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "jitterentropy": [{"name": "jitterentropy", "version": "3.5.0", "release": "4.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "rng-tools": [{"name": "rng-tools", "version": "6.17", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}], "python3-pip": [{"name": "python3-pip", "version": "23.3.2", "release": "3.el10", "epoch": null, "arch": "noarch", "source": "rpm"}], "dnsmasq": [{"name": "dnsmasq", "version": "2.90", "release": "3.el10", "epoch": null, "arch": "x86_64", "source": "rpm"}]}}, "invocation": {"module_args": {"manager": ["auto"], "strategy": "first"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203884.16185: done with _execute_module (package_facts, {'_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'package_facts', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203884.16189: _low_level_execute_command(): starting 13731 1727203884.16192: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203883.4910989-18164-6394255790301/ > /dev/null 2>&1 && sleep 0' 13731 1727203884.17481: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203884.17485: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203884.17504: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203884.17512: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203884.17527: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203884.17853: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203884.17857: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203884.17860: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203884.17883: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203884.17940: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203884.19791: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203884.19795: stderr chunk (state=3): >>><<< 13731 1727203884.19798: stdout chunk (state=3): >>><<< 13731 1727203884.19817: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203884.19823: handler run complete 13731 1727203884.21671: variable 'ansible_facts' from source: unknown 13731 1727203884.22915: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.26880: variable 'ansible_facts' from source: unknown 13731 1727203884.27965: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.29591: attempt loop complete, returning result 13731 1727203884.29594: _execute() done 13731 1727203884.29596: dumping result to json 13731 1727203884.29910: done dumping result, returning 13731 1727203884.29919: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check which packages are installed [028d2410-947f-82dc-c122-000000000fe6] 13731 1727203884.30041: sending task result for task 028d2410-947f-82dc-c122-000000000fe6 13731 1727203884.33858: done sending task result for task 028d2410-947f-82dc-c122-000000000fe6 13731 1727203884.33862: WORKER PROCESS EXITING ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203884.34014: no more pending results, returning what we have 13731 1727203884.34018: results queue empty 13731 1727203884.34018: checking for any_errors_fatal 13731 1727203884.34023: done checking for any_errors_fatal 13731 1727203884.34024: checking for max_fail_percentage 13731 1727203884.34030: done checking for max_fail_percentage 13731 1727203884.34031: checking to see if all hosts have failed and the running result is not ok 13731 1727203884.34031: done checking to see if all hosts have failed 13731 1727203884.34032: getting the remaining hosts for this loop 13731 1727203884.34034: done getting the remaining hosts for this loop 13731 1727203884.34037: getting the next task for host managed-node3 13731 1727203884.34044: done getting next task for host managed-node3 13731 1727203884.34048: ^ task is: TASK: fedora.linux_system_roles.network : Print network provider 13731 1727203884.34054: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203884.34065: getting variables 13731 1727203884.34066: in VariableManager get_vars() 13731 1727203884.34102: Calling all_inventory to load vars for managed-node3 13731 1727203884.34105: Calling groups_inventory to load vars for managed-node3 13731 1727203884.34107: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203884.34116: Calling all_plugins_play to load vars for managed-node3 13731 1727203884.34118: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203884.34121: Calling groups_plugins_play to load vars for managed-node3 13731 1727203884.35635: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.37388: done with get_vars() 13731 1727203884.37416: done getting variables 13731 1727203884.37491: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Print network provider] ************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:7 Tuesday 24 September 2024 14:51:24 -0400 (0:00:00.937) 0:01:02.606 ***** 13731 1727203884.37530: entering _queue_task() for managed-node3/debug 13731 1727203884.38110: worker is 1 (out of 1 available) 13731 1727203884.38122: exiting _queue_task() for managed-node3/debug 13731 1727203884.38134: done queuing things up, now waiting for results queue to drain 13731 1727203884.38135: waiting for pending results... 13731 1727203884.38506: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Print network provider 13731 1727203884.38668: in run() - task 028d2410-947f-82dc-c122-000000000e0d 13731 1727203884.38693: variable 'ansible_search_path' from source: unknown 13731 1727203884.38700: variable 'ansible_search_path' from source: unknown 13731 1727203884.38743: calling self._execute() 13731 1727203884.38847: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.38867: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.38886: variable 'omit' from source: magic vars 13731 1727203884.39278: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.39400: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203884.39405: variable 'omit' from source: magic vars 13731 1727203884.39408: variable 'omit' from source: magic vars 13731 1727203884.39510: variable 'network_provider' from source: set_fact 13731 1727203884.39534: variable 'omit' from source: magic vars 13731 1727203884.39584: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203884.39756: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203884.39786: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203884.39809: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203884.39829: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203884.39871: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203884.39945: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.39949: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.40004: Set connection var ansible_pipelining to False 13731 1727203884.40016: Set connection var ansible_shell_type to sh 13731 1727203884.40026: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203884.40035: Set connection var ansible_connection to ssh 13731 1727203884.40045: Set connection var ansible_shell_executable to /bin/sh 13731 1727203884.40067: Set connection var ansible_timeout to 10 13731 1727203884.40094: variable 'ansible_shell_executable' from source: unknown 13731 1727203884.40103: variable 'ansible_connection' from source: unknown 13731 1727203884.40110: variable 'ansible_module_compression' from source: unknown 13731 1727203884.40117: variable 'ansible_shell_type' from source: unknown 13731 1727203884.40167: variable 'ansible_shell_executable' from source: unknown 13731 1727203884.40170: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.40172: variable 'ansible_pipelining' from source: unknown 13731 1727203884.40174: variable 'ansible_timeout' from source: unknown 13731 1727203884.40178: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.40303: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203884.40317: variable 'omit' from source: magic vars 13731 1727203884.40326: starting attempt loop 13731 1727203884.40332: running the handler 13731 1727203884.40389: handler run complete 13731 1727203884.40407: attempt loop complete, returning result 13731 1727203884.40482: _execute() done 13731 1727203884.40487: dumping result to json 13731 1727203884.40489: done dumping result, returning 13731 1727203884.40491: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Print network provider [028d2410-947f-82dc-c122-000000000e0d] 13731 1727203884.40493: sending task result for task 028d2410-947f-82dc-c122-000000000e0d 13731 1727203884.40556: done sending task result for task 028d2410-947f-82dc-c122-000000000e0d 13731 1727203884.40559: WORKER PROCESS EXITING ok: [managed-node3] => {} MSG: Using network provider: nm 13731 1727203884.40629: no more pending results, returning what we have 13731 1727203884.40633: results queue empty 13731 1727203884.40634: checking for any_errors_fatal 13731 1727203884.40647: done checking for any_errors_fatal 13731 1727203884.40648: checking for max_fail_percentage 13731 1727203884.40650: done checking for max_fail_percentage 13731 1727203884.40651: checking to see if all hosts have failed and the running result is not ok 13731 1727203884.40651: done checking to see if all hosts have failed 13731 1727203884.40652: getting the remaining hosts for this loop 13731 1727203884.40655: done getting the remaining hosts for this loop 13731 1727203884.40659: getting the next task for host managed-node3 13731 1727203884.40670: done getting next task for host managed-node3 13731 1727203884.40674: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13731 1727203884.40683: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203884.40696: getting variables 13731 1727203884.40698: in VariableManager get_vars() 13731 1727203884.40746: Calling all_inventory to load vars for managed-node3 13731 1727203884.40749: Calling groups_inventory to load vars for managed-node3 13731 1727203884.40751: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203884.40764: Calling all_plugins_play to load vars for managed-node3 13731 1727203884.40767: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203884.40771: Calling groups_plugins_play to load vars for managed-node3 13731 1727203884.42526: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.45068: done with get_vars() 13731 1727203884.45101: done getting variables 13731 1727203884.45163: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:11 Tuesday 24 September 2024 14:51:24 -0400 (0:00:00.076) 0:01:02.683 ***** 13731 1727203884.45211: entering _queue_task() for managed-node3/fail 13731 1727203884.45554: worker is 1 (out of 1 available) 13731 1727203884.45569: exiting _queue_task() for managed-node3/fail 13731 1727203884.45585: done queuing things up, now waiting for results queue to drain 13731 1727203884.45586: waiting for pending results... 13731 1727203884.46111: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider 13731 1727203884.46493: in run() - task 028d2410-947f-82dc-c122-000000000e0e 13731 1727203884.46497: variable 'ansible_search_path' from source: unknown 13731 1727203884.46500: variable 'ansible_search_path' from source: unknown 13731 1727203884.46502: calling self._execute() 13731 1727203884.46636: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.46648: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.46668: variable 'omit' from source: magic vars 13731 1727203884.47467: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.47545: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203884.47977: variable 'network_state' from source: role '' defaults 13731 1727203884.47981: Evaluated conditional (network_state != {}): False 13731 1727203884.47984: when evaluation is False, skipping this task 13731 1727203884.47986: _execute() done 13731 1727203884.47988: dumping result to json 13731 1727203884.47990: done dumping result, returning 13731 1727203884.47992: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if using the `network_state` variable with the initscripts provider [028d2410-947f-82dc-c122-000000000e0e] 13731 1727203884.47996: sending task result for task 028d2410-947f-82dc-c122-000000000e0e 13731 1727203884.48067: done sending task result for task 028d2410-947f-82dc-c122-000000000e0e 13731 1727203884.48070: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203884.48236: no more pending results, returning what we have 13731 1727203884.48240: results queue empty 13731 1727203884.48241: checking for any_errors_fatal 13731 1727203884.48248: done checking for any_errors_fatal 13731 1727203884.48248: checking for max_fail_percentage 13731 1727203884.48250: done checking for max_fail_percentage 13731 1727203884.48251: checking to see if all hosts have failed and the running result is not ok 13731 1727203884.48252: done checking to see if all hosts have failed 13731 1727203884.48256: getting the remaining hosts for this loop 13731 1727203884.48261: done getting the remaining hosts for this loop 13731 1727203884.48266: getting the next task for host managed-node3 13731 1727203884.48277: done getting next task for host managed-node3 13731 1727203884.48282: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13731 1727203884.48289: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203884.48313: getting variables 13731 1727203884.48315: in VariableManager get_vars() 13731 1727203884.48358: Calling all_inventory to load vars for managed-node3 13731 1727203884.48364: Calling groups_inventory to load vars for managed-node3 13731 1727203884.48366: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203884.48484: Calling all_plugins_play to load vars for managed-node3 13731 1727203884.48488: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203884.48491: Calling groups_plugins_play to load vars for managed-node3 13731 1727203884.49855: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.52455: done with get_vars() 13731 1727203884.52489: done getting variables 13731 1727203884.52553: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:18 Tuesday 24 September 2024 14:51:24 -0400 (0:00:00.073) 0:01:02.756 ***** 13731 1727203884.52596: entering _queue_task() for managed-node3/fail 13731 1727203884.52947: worker is 1 (out of 1 available) 13731 1727203884.52961: exiting _queue_task() for managed-node3/fail 13731 1727203884.52973: done queuing things up, now waiting for results queue to drain 13731 1727203884.52975: waiting for pending results... 13731 1727203884.53219: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 13731 1727203884.53474: in run() - task 028d2410-947f-82dc-c122-000000000e0f 13731 1727203884.53480: variable 'ansible_search_path' from source: unknown 13731 1727203884.53484: variable 'ansible_search_path' from source: unknown 13731 1727203884.53486: calling self._execute() 13731 1727203884.53559: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.53579: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.53597: variable 'omit' from source: magic vars 13731 1727203884.53985: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.54001: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203884.54134: variable 'network_state' from source: role '' defaults 13731 1727203884.54149: Evaluated conditional (network_state != {}): False 13731 1727203884.54157: when evaluation is False, skipping this task 13731 1727203884.54164: _execute() done 13731 1727203884.54172: dumping result to json 13731 1727203884.54235: done dumping result, returning 13731 1727203884.54239: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying the network state configuration if the system version of the managed host is below 8 [028d2410-947f-82dc-c122-000000000e0f] 13731 1727203884.54242: sending task result for task 028d2410-947f-82dc-c122-000000000e0f 13731 1727203884.54314: done sending task result for task 028d2410-947f-82dc-c122-000000000e0f 13731 1727203884.54317: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203884.54388: no more pending results, returning what we have 13731 1727203884.54392: results queue empty 13731 1727203884.54393: checking for any_errors_fatal 13731 1727203884.54403: done checking for any_errors_fatal 13731 1727203884.54404: checking for max_fail_percentage 13731 1727203884.54406: done checking for max_fail_percentage 13731 1727203884.54407: checking to see if all hosts have failed and the running result is not ok 13731 1727203884.54408: done checking to see if all hosts have failed 13731 1727203884.54409: getting the remaining hosts for this loop 13731 1727203884.54411: done getting the remaining hosts for this loop 13731 1727203884.54415: getting the next task for host managed-node3 13731 1727203884.54423: done getting next task for host managed-node3 13731 1727203884.54428: ^ task is: TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13731 1727203884.54435: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203884.54466: getting variables 13731 1727203884.54468: in VariableManager get_vars() 13731 1727203884.54517: Calling all_inventory to load vars for managed-node3 13731 1727203884.54521: Calling groups_inventory to load vars for managed-node3 13731 1727203884.54523: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203884.54536: Calling all_plugins_play to load vars for managed-node3 13731 1727203884.54539: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203884.54542: Calling groups_plugins_play to load vars for managed-node3 13731 1727203884.56342: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.57942: done with get_vars() 13731 1727203884.57966: done getting variables 13731 1727203884.58029: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:25 Tuesday 24 September 2024 14:51:24 -0400 (0:00:00.054) 0:01:02.811 ***** 13731 1727203884.58065: entering _queue_task() for managed-node3/fail 13731 1727203884.58398: worker is 1 (out of 1 available) 13731 1727203884.58410: exiting _queue_task() for managed-node3/fail 13731 1727203884.58426: done queuing things up, now waiting for results queue to drain 13731 1727203884.58428: waiting for pending results... 13731 1727203884.58659: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later 13731 1727203884.58863: in run() - task 028d2410-947f-82dc-c122-000000000e10 13731 1727203884.58868: variable 'ansible_search_path' from source: unknown 13731 1727203884.58871: variable 'ansible_search_path' from source: unknown 13731 1727203884.58874: calling self._execute() 13731 1727203884.58972: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.58978: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.58985: variable 'omit' from source: magic vars 13731 1727203884.59426: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.59438: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203884.59633: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203884.61895: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203884.62026: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203884.62030: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203884.62033: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203884.62056: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203884.62138: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.62167: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.62197: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.62251: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.62259: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.62366: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.62372: Evaluated conditional (ansible_distribution_major_version | int > 9): True 13731 1727203884.62478: variable 'ansible_distribution' from source: facts 13731 1727203884.62567: variable '__network_rh_distros' from source: role '' defaults 13731 1727203884.62571: Evaluated conditional (ansible_distribution in __network_rh_distros): True 13731 1727203884.62739: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.62762: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.62790: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.62829: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.62846: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.62899: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.62980: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.62984: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.62986: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.62999: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.63038: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.63067: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.63092: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.63127: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.63141: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.63495: variable 'network_connections' from source: task vars 13731 1727203884.63506: variable 'port2_profile' from source: play vars 13731 1727203884.63567: variable 'port2_profile' from source: play vars 13731 1727203884.63661: variable 'port1_profile' from source: play vars 13731 1727203884.63665: variable 'port1_profile' from source: play vars 13731 1727203884.63667: variable 'controller_profile' from source: play vars 13731 1727203884.63717: variable 'controller_profile' from source: play vars 13731 1727203884.63725: variable 'network_state' from source: role '' defaults 13731 1727203884.63800: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203884.63989: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203884.64024: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203884.64057: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203884.64094: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203884.64284: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203884.64287: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203884.64289: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.64291: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203884.64294: Evaluated conditional (network_connections | selectattr("type", "defined") | selectattr("type", "match", "^team$") | list | length > 0 or network_state.get("interfaces", []) | selectattr("type", "defined") | selectattr("type", "match", "^team$") | list | length > 0): False 13731 1727203884.64296: when evaluation is False, skipping this task 13731 1727203884.64298: _execute() done 13731 1727203884.64300: dumping result to json 13731 1727203884.64302: done dumping result, returning 13731 1727203884.64304: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Abort applying teaming configuration if the system version of the managed host is EL10 or later [028d2410-947f-82dc-c122-000000000e10] 13731 1727203884.64307: sending task result for task 028d2410-947f-82dc-c122-000000000e10 13731 1727203884.64370: done sending task result for task 028d2410-947f-82dc-c122-000000000e10 13731 1727203884.64373: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_connections | selectattr(\"type\", \"defined\") | selectattr(\"type\", \"match\", \"^team$\") | list | length > 0 or network_state.get(\"interfaces\", []) | selectattr(\"type\", \"defined\") | selectattr(\"type\", \"match\", \"^team$\") | list | length > 0", "skip_reason": "Conditional result was False" } 13731 1727203884.64418: no more pending results, returning what we have 13731 1727203884.64422: results queue empty 13731 1727203884.64423: checking for any_errors_fatal 13731 1727203884.64431: done checking for any_errors_fatal 13731 1727203884.64432: checking for max_fail_percentage 13731 1727203884.64434: done checking for max_fail_percentage 13731 1727203884.64434: checking to see if all hosts have failed and the running result is not ok 13731 1727203884.64435: done checking to see if all hosts have failed 13731 1727203884.64436: getting the remaining hosts for this loop 13731 1727203884.64438: done getting the remaining hosts for this loop 13731 1727203884.64441: getting the next task for host managed-node3 13731 1727203884.64449: done getting next task for host managed-node3 13731 1727203884.64452: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13731 1727203884.64457: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203884.64480: getting variables 13731 1727203884.64481: in VariableManager get_vars() 13731 1727203884.64525: Calling all_inventory to load vars for managed-node3 13731 1727203884.64528: Calling groups_inventory to load vars for managed-node3 13731 1727203884.64530: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203884.64540: Calling all_plugins_play to load vars for managed-node3 13731 1727203884.64542: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203884.64545: Calling groups_plugins_play to load vars for managed-node3 13731 1727203884.66731: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.69196: done with get_vars() 13731 1727203884.69231: done getting variables 13731 1727203884.69303: Loading ActionModule 'dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:36 Tuesday 24 September 2024 14:51:24 -0400 (0:00:00.112) 0:01:02.924 ***** 13731 1727203884.69345: entering _queue_task() for managed-node3/dnf 13731 1727203884.69735: worker is 1 (out of 1 available) 13731 1727203884.69749: exiting _queue_task() for managed-node3/dnf 13731 1727203884.69764: done queuing things up, now waiting for results queue to drain 13731 1727203884.69766: waiting for pending results... 13731 1727203884.70134: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces 13731 1727203884.70384: in run() - task 028d2410-947f-82dc-c122-000000000e11 13731 1727203884.70389: variable 'ansible_search_path' from source: unknown 13731 1727203884.70392: variable 'ansible_search_path' from source: unknown 13731 1727203884.70395: calling self._execute() 13731 1727203884.70551: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.70731: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.70735: variable 'omit' from source: magic vars 13731 1727203884.71107: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.71125: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203884.71347: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203884.73670: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203884.73724: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203884.73754: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203884.73778: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203884.73800: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203884.73857: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.73881: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.73899: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.73927: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.73942: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.74059: variable 'ansible_distribution' from source: facts 13731 1727203884.74065: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.74134: Evaluated conditional (ansible_distribution == 'Fedora' or ansible_distribution_major_version | int > 7): True 13731 1727203884.74304: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203884.74440: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.74479: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.74510: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.74605: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.74609: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.74648: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.74684: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.74714: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.74850: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.74853: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.74855: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.74866: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.74897: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.74938: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.74969: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.75145: variable 'network_connections' from source: task vars 13731 1727203884.75179: variable 'port2_profile' from source: play vars 13731 1727203884.75281: variable 'port2_profile' from source: play vars 13731 1727203884.75285: variable 'port1_profile' from source: play vars 13731 1727203884.75345: variable 'port1_profile' from source: play vars 13731 1727203884.75348: variable 'controller_profile' from source: play vars 13731 1727203884.75400: variable 'controller_profile' from source: play vars 13731 1727203884.75447: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203884.75571: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203884.75601: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203884.75635: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203884.75656: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203884.75689: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203884.75707: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203884.75729: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.75747: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203884.75784: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203884.75935: variable 'network_connections' from source: task vars 13731 1727203884.75939: variable 'port2_profile' from source: play vars 13731 1727203884.75982: variable 'port2_profile' from source: play vars 13731 1727203884.75989: variable 'port1_profile' from source: play vars 13731 1727203884.76030: variable 'port1_profile' from source: play vars 13731 1727203884.76037: variable 'controller_profile' from source: play vars 13731 1727203884.76083: variable 'controller_profile' from source: play vars 13731 1727203884.76100: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203884.76103: when evaluation is False, skipping this task 13731 1727203884.76106: _execute() done 13731 1727203884.76108: dumping result to json 13731 1727203884.76110: done dumping result, returning 13731 1727203884.76118: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the DNF package manager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000e11] 13731 1727203884.76123: sending task result for task 028d2410-947f-82dc-c122-000000000e11 13731 1727203884.76212: done sending task result for task 028d2410-947f-82dc-c122-000000000e11 13731 1727203884.76215: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203884.76270: no more pending results, returning what we have 13731 1727203884.76274: results queue empty 13731 1727203884.76275: checking for any_errors_fatal 13731 1727203884.76283: done checking for any_errors_fatal 13731 1727203884.76284: checking for max_fail_percentage 13731 1727203884.76285: done checking for max_fail_percentage 13731 1727203884.76286: checking to see if all hosts have failed and the running result is not ok 13731 1727203884.76287: done checking to see if all hosts have failed 13731 1727203884.76288: getting the remaining hosts for this loop 13731 1727203884.76289: done getting the remaining hosts for this loop 13731 1727203884.76293: getting the next task for host managed-node3 13731 1727203884.76301: done getting next task for host managed-node3 13731 1727203884.76305: ^ task is: TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13731 1727203884.76310: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203884.76336: getting variables 13731 1727203884.76338: in VariableManager get_vars() 13731 1727203884.76384: Calling all_inventory to load vars for managed-node3 13731 1727203884.76387: Calling groups_inventory to load vars for managed-node3 13731 1727203884.76389: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203884.76397: Calling all_plugins_play to load vars for managed-node3 13731 1727203884.76400: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203884.76402: Calling groups_plugins_play to load vars for managed-node3 13731 1727203884.77387: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.78753: done with get_vars() 13731 1727203884.78781: done getting variables redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf 13731 1727203884.78855: Loading ActionModule 'ansible_collections.ansible.builtin.plugins.action.dnf' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/dnf.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:48 Tuesday 24 September 2024 14:51:24 -0400 (0:00:00.095) 0:01:03.019 ***** 13731 1727203884.78893: entering _queue_task() for managed-node3/yum 13731 1727203884.79219: worker is 1 (out of 1 available) 13731 1727203884.79231: exiting _queue_task() for managed-node3/yum 13731 1727203884.79244: done queuing things up, now waiting for results queue to drain 13731 1727203884.79246: waiting for pending results... 13731 1727203884.79560: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces 13731 1727203884.79654: in run() - task 028d2410-947f-82dc-c122-000000000e12 13731 1727203884.79667: variable 'ansible_search_path' from source: unknown 13731 1727203884.79673: variable 'ansible_search_path' from source: unknown 13731 1727203884.79706: calling self._execute() 13731 1727203884.79788: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.79791: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.79799: variable 'omit' from source: magic vars 13731 1727203884.80080: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.80091: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203884.80216: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203884.82081: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203884.82085: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203884.82087: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203884.82111: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203884.82143: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203884.82235: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.82269: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.82330: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.82382: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.82394: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.82493: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.82507: Evaluated conditional (ansible_distribution_major_version | int < 8): False 13731 1727203884.82510: when evaluation is False, skipping this task 13731 1727203884.82513: _execute() done 13731 1727203884.82515: dumping result to json 13731 1727203884.82517: done dumping result, returning 13731 1727203884.82525: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Check if updates for network packages are available through the YUM package manager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000e12] 13731 1727203884.82538: sending task result for task 028d2410-947f-82dc-c122-000000000e12 13731 1727203884.82620: done sending task result for task 028d2410-947f-82dc-c122-000000000e12 13731 1727203884.82623: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "ansible_distribution_major_version | int < 8", "skip_reason": "Conditional result was False" } 13731 1727203884.82686: no more pending results, returning what we have 13731 1727203884.82689: results queue empty 13731 1727203884.82690: checking for any_errors_fatal 13731 1727203884.82696: done checking for any_errors_fatal 13731 1727203884.82696: checking for max_fail_percentage 13731 1727203884.82698: done checking for max_fail_percentage 13731 1727203884.82699: checking to see if all hosts have failed and the running result is not ok 13731 1727203884.82700: done checking to see if all hosts have failed 13731 1727203884.82700: getting the remaining hosts for this loop 13731 1727203884.82702: done getting the remaining hosts for this loop 13731 1727203884.82705: getting the next task for host managed-node3 13731 1727203884.82714: done getting next task for host managed-node3 13731 1727203884.82718: ^ task is: TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13731 1727203884.82723: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203884.82746: getting variables 13731 1727203884.82748: in VariableManager get_vars() 13731 1727203884.82793: Calling all_inventory to load vars for managed-node3 13731 1727203884.82796: Calling groups_inventory to load vars for managed-node3 13731 1727203884.82798: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203884.82807: Calling all_plugins_play to load vars for managed-node3 13731 1727203884.82809: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203884.82812: Calling groups_plugins_play to load vars for managed-node3 13731 1727203884.84153: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.85128: done with get_vars() 13731 1727203884.85146: done getting variables 13731 1727203884.85191: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:60 Tuesday 24 September 2024 14:51:24 -0400 (0:00:00.063) 0:01:03.083 ***** 13731 1727203884.85217: entering _queue_task() for managed-node3/fail 13731 1727203884.85449: worker is 1 (out of 1 available) 13731 1727203884.85466: exiting _queue_task() for managed-node3/fail 13731 1727203884.85479: done queuing things up, now waiting for results queue to drain 13731 1727203884.85481: waiting for pending results... 13731 1727203884.85663: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces 13731 1727203884.85751: in run() - task 028d2410-947f-82dc-c122-000000000e13 13731 1727203884.85765: variable 'ansible_search_path' from source: unknown 13731 1727203884.85770: variable 'ansible_search_path' from source: unknown 13731 1727203884.85797: calling self._execute() 13731 1727203884.85873: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.85878: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.85887: variable 'omit' from source: magic vars 13731 1727203884.86481: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.86485: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203884.86588: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203884.86918: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203884.89248: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203884.89320: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203884.89369: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203884.89410: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203884.89680: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203884.89684: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.89687: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.89690: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.89692: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.89694: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.89696: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.89725: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.89754: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.89801: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.89829: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.89876: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203884.89901: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203884.89932: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.89969: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203884.89987: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203884.90177: variable 'network_connections' from source: task vars 13731 1727203884.90197: variable 'port2_profile' from source: play vars 13731 1727203884.90282: variable 'port2_profile' from source: play vars 13731 1727203884.90299: variable 'port1_profile' from source: play vars 13731 1727203884.90374: variable 'port1_profile' from source: play vars 13731 1727203884.90390: variable 'controller_profile' from source: play vars 13731 1727203884.90452: variable 'controller_profile' from source: play vars 13731 1727203884.90534: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203884.90736: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203884.90782: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203884.90825: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203884.90857: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203884.90917: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203884.90944: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203884.90979: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203884.91080: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203884.91084: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203884.91335: variable 'network_connections' from source: task vars 13731 1727203884.91346: variable 'port2_profile' from source: play vars 13731 1727203884.91412: variable 'port2_profile' from source: play vars 13731 1727203884.91424: variable 'port1_profile' from source: play vars 13731 1727203884.91496: variable 'port1_profile' from source: play vars 13731 1727203884.91509: variable 'controller_profile' from source: play vars 13731 1727203884.91580: variable 'controller_profile' from source: play vars 13731 1727203884.91610: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203884.91627: when evaluation is False, skipping this task 13731 1727203884.91666: _execute() done 13731 1727203884.91670: dumping result to json 13731 1727203884.91672: done dumping result, returning 13731 1727203884.91676: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ask user's consent to restart NetworkManager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000e13] 13731 1727203884.91679: sending task result for task 028d2410-947f-82dc-c122-000000000e13 skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203884.91954: no more pending results, returning what we have 13731 1727203884.91958: results queue empty 13731 1727203884.91959: checking for any_errors_fatal 13731 1727203884.91970: done checking for any_errors_fatal 13731 1727203884.91971: checking for max_fail_percentage 13731 1727203884.91973: done checking for max_fail_percentage 13731 1727203884.91974: checking to see if all hosts have failed and the running result is not ok 13731 1727203884.91976: done checking to see if all hosts have failed 13731 1727203884.91977: getting the remaining hosts for this loop 13731 1727203884.91979: done getting the remaining hosts for this loop 13731 1727203884.91988: getting the next task for host managed-node3 13731 1727203884.91998: done getting next task for host managed-node3 13731 1727203884.92002: ^ task is: TASK: fedora.linux_system_roles.network : Install packages 13731 1727203884.92007: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203884.92098: getting variables 13731 1727203884.92101: in VariableManager get_vars() 13731 1727203884.92149: Calling all_inventory to load vars for managed-node3 13731 1727203884.92152: Calling groups_inventory to load vars for managed-node3 13731 1727203884.92155: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203884.92168: Calling all_plugins_play to load vars for managed-node3 13731 1727203884.92172: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203884.92315: Calling groups_plugins_play to load vars for managed-node3 13731 1727203884.92988: done sending task result for task 028d2410-947f-82dc-c122-000000000e13 13731 1727203884.92991: WORKER PROCESS EXITING 13731 1727203884.93735: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203884.95323: done with get_vars() 13731 1727203884.95348: done getting variables 13731 1727203884.95417: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install packages] ******************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:73 Tuesday 24 September 2024 14:51:24 -0400 (0:00:00.102) 0:01:03.185 ***** 13731 1727203884.95456: entering _queue_task() for managed-node3/package 13731 1727203884.95820: worker is 1 (out of 1 available) 13731 1727203884.95949: exiting _queue_task() for managed-node3/package 13731 1727203884.95959: done queuing things up, now waiting for results queue to drain 13731 1727203884.95963: waiting for pending results... 13731 1727203884.96165: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install packages 13731 1727203884.96337: in run() - task 028d2410-947f-82dc-c122-000000000e14 13731 1727203884.96355: variable 'ansible_search_path' from source: unknown 13731 1727203884.96364: variable 'ansible_search_path' from source: unknown 13731 1727203884.96410: calling self._execute() 13731 1727203884.96517: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203884.96528: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203884.96541: variable 'omit' from source: magic vars 13731 1727203884.96938: variable 'ansible_distribution_major_version' from source: facts 13731 1727203884.96954: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203884.97168: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203884.97446: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203884.97583: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203884.97587: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203884.97589: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203884.97694: variable 'network_packages' from source: role '' defaults 13731 1727203884.97817: variable '__network_provider_setup' from source: role '' defaults 13731 1727203884.97830: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203884.97898: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203884.97920: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203884.98029: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203884.98232: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203885.00302: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203885.00742: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203885.00880: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203885.00884: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203885.00886: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203885.00943: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.00982: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.01022: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.01071: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.01094: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.01221: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.01224: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.01227: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.01253: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.01279: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.01499: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13731 1727203885.01614: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.01641: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.01680: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.01721: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.01737: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.01832: variable 'ansible_python' from source: facts 13731 1727203885.01854: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13731 1727203885.01947: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203885.02091: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203885.02178: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.02213: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.02240: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.02288: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.02313: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.02360: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.02401: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.02586: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.02589: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.02592: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.02649: variable 'network_connections' from source: task vars 13731 1727203885.02663: variable 'port2_profile' from source: play vars 13731 1727203885.02773: variable 'port2_profile' from source: play vars 13731 1727203885.02791: variable 'port1_profile' from source: play vars 13731 1727203885.02899: variable 'port1_profile' from source: play vars 13731 1727203885.02913: variable 'controller_profile' from source: play vars 13731 1727203885.03020: variable 'controller_profile' from source: play vars 13731 1727203885.03104: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203885.03133: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203885.03178: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.03213: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203885.03273: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203885.03569: variable 'network_connections' from source: task vars 13731 1727203885.03686: variable 'port2_profile' from source: play vars 13731 1727203885.03692: variable 'port2_profile' from source: play vars 13731 1727203885.03706: variable 'port1_profile' from source: play vars 13731 1727203885.03814: variable 'port1_profile' from source: play vars 13731 1727203885.03828: variable 'controller_profile' from source: play vars 13731 1727203885.03937: variable 'controller_profile' from source: play vars 13731 1727203885.03975: variable '__network_packages_default_wireless' from source: role '' defaults 13731 1727203885.04068: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203885.04402: variable 'network_connections' from source: task vars 13731 1727203885.04446: variable 'port2_profile' from source: play vars 13731 1727203885.04491: variable 'port2_profile' from source: play vars 13731 1727203885.04503: variable 'port1_profile' from source: play vars 13731 1727203885.04579: variable 'port1_profile' from source: play vars 13731 1727203885.04593: variable 'controller_profile' from source: play vars 13731 1727203885.04648: variable 'controller_profile' from source: play vars 13731 1727203885.04773: variable '__network_packages_default_team' from source: role '' defaults 13731 1727203885.04780: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203885.05097: variable 'network_connections' from source: task vars 13731 1727203885.05113: variable 'port2_profile' from source: play vars 13731 1727203885.05216: variable 'port2_profile' from source: play vars 13731 1727203885.05219: variable 'port1_profile' from source: play vars 13731 1727203885.05267: variable 'port1_profile' from source: play vars 13731 1727203885.05283: variable 'controller_profile' from source: play vars 13731 1727203885.05356: variable 'controller_profile' from source: play vars 13731 1727203885.05433: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203885.05496: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203885.05543: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203885.05583: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203885.05811: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13731 1727203885.12832: variable 'network_connections' from source: task vars 13731 1727203885.12844: variable 'port2_profile' from source: play vars 13731 1727203885.12981: variable 'port2_profile' from source: play vars 13731 1727203885.12984: variable 'port1_profile' from source: play vars 13731 1727203885.12991: variable 'port1_profile' from source: play vars 13731 1727203885.13012: variable 'controller_profile' from source: play vars 13731 1727203885.13079: variable 'controller_profile' from source: play vars 13731 1727203885.13094: variable 'ansible_distribution' from source: facts 13731 1727203885.13109: variable '__network_rh_distros' from source: role '' defaults 13731 1727203885.13181: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.13184: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13731 1727203885.13328: variable 'ansible_distribution' from source: facts 13731 1727203885.13338: variable '__network_rh_distros' from source: role '' defaults 13731 1727203885.13349: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.13367: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13731 1727203885.13543: variable 'ansible_distribution' from source: facts 13731 1727203885.13556: variable '__network_rh_distros' from source: role '' defaults 13731 1727203885.13655: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.13658: variable 'network_provider' from source: set_fact 13731 1727203885.13664: variable 'ansible_facts' from source: unknown 13731 1727203885.14277: Evaluated conditional (not network_packages is subset(ansible_facts.packages.keys())): False 13731 1727203885.14286: when evaluation is False, skipping this task 13731 1727203885.14294: _execute() done 13731 1727203885.14305: dumping result to json 13731 1727203885.14317: done dumping result, returning 13731 1727203885.14329: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install packages [028d2410-947f-82dc-c122-000000000e14] 13731 1727203885.14337: sending task result for task 028d2410-947f-82dc-c122-000000000e14 skipping: [managed-node3] => { "changed": false, "false_condition": "not network_packages is subset(ansible_facts.packages.keys())", "skip_reason": "Conditional result was False" } 13731 1727203885.14493: no more pending results, returning what we have 13731 1727203885.14497: results queue empty 13731 1727203885.14498: checking for any_errors_fatal 13731 1727203885.14505: done checking for any_errors_fatal 13731 1727203885.14506: checking for max_fail_percentage 13731 1727203885.14508: done checking for max_fail_percentage 13731 1727203885.14509: checking to see if all hosts have failed and the running result is not ok 13731 1727203885.14510: done checking to see if all hosts have failed 13731 1727203885.14510: getting the remaining hosts for this loop 13731 1727203885.14512: done getting the remaining hosts for this loop 13731 1727203885.14516: getting the next task for host managed-node3 13731 1727203885.14524: done getting next task for host managed-node3 13731 1727203885.14534: ^ task is: TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13731 1727203885.14540: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203885.14563: getting variables 13731 1727203885.14565: in VariableManager get_vars() 13731 1727203885.14730: Calling all_inventory to load vars for managed-node3 13731 1727203885.14733: Calling groups_inventory to load vars for managed-node3 13731 1727203885.14736: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203885.14745: Calling all_plugins_play to load vars for managed-node3 13731 1727203885.14748: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203885.14752: Calling groups_plugins_play to load vars for managed-node3 13731 1727203885.15391: done sending task result for task 028d2410-947f-82dc-c122-000000000e14 13731 1727203885.15394: WORKER PROCESS EXITING 13731 1727203885.21870: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203885.23142: done with get_vars() 13731 1727203885.23166: done getting variables 13731 1727203885.23202: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:85 Tuesday 24 September 2024 14:51:25 -0400 (0:00:00.277) 0:01:03.463 ***** 13731 1727203885.23225: entering _queue_task() for managed-node3/package 13731 1727203885.23496: worker is 1 (out of 1 available) 13731 1727203885.23510: exiting _queue_task() for managed-node3/package 13731 1727203885.23522: done queuing things up, now waiting for results queue to drain 13731 1727203885.23525: waiting for pending results... 13731 1727203885.23721: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable 13731 1727203885.23838: in run() - task 028d2410-947f-82dc-c122-000000000e15 13731 1727203885.23849: variable 'ansible_search_path' from source: unknown 13731 1727203885.23854: variable 'ansible_search_path' from source: unknown 13731 1727203885.23889: calling self._execute() 13731 1727203885.23961: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203885.23973: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203885.23979: variable 'omit' from source: magic vars 13731 1727203885.24262: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.24274: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203885.24361: variable 'network_state' from source: role '' defaults 13731 1727203885.24373: Evaluated conditional (network_state != {}): False 13731 1727203885.24378: when evaluation is False, skipping this task 13731 1727203885.24381: _execute() done 13731 1727203885.24383: dumping result to json 13731 1727203885.24386: done dumping result, returning 13731 1727203885.24393: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install NetworkManager and nmstate when using network_state variable [028d2410-947f-82dc-c122-000000000e15] 13731 1727203885.24399: sending task result for task 028d2410-947f-82dc-c122-000000000e15 13731 1727203885.24517: done sending task result for task 028d2410-947f-82dc-c122-000000000e15 13731 1727203885.24520: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203885.24599: no more pending results, returning what we have 13731 1727203885.24603: results queue empty 13731 1727203885.24604: checking for any_errors_fatal 13731 1727203885.24612: done checking for any_errors_fatal 13731 1727203885.24613: checking for max_fail_percentage 13731 1727203885.24614: done checking for max_fail_percentage 13731 1727203885.24615: checking to see if all hosts have failed and the running result is not ok 13731 1727203885.24616: done checking to see if all hosts have failed 13731 1727203885.24616: getting the remaining hosts for this loop 13731 1727203885.24618: done getting the remaining hosts for this loop 13731 1727203885.24621: getting the next task for host managed-node3 13731 1727203885.24630: done getting next task for host managed-node3 13731 1727203885.24634: ^ task is: TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13731 1727203885.24640: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203885.24667: getting variables 13731 1727203885.24668: in VariableManager get_vars() 13731 1727203885.24714: Calling all_inventory to load vars for managed-node3 13731 1727203885.24718: Calling groups_inventory to load vars for managed-node3 13731 1727203885.24720: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203885.24731: Calling all_plugins_play to load vars for managed-node3 13731 1727203885.24733: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203885.24737: Calling groups_plugins_play to load vars for managed-node3 13731 1727203885.26048: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203885.26934: done with get_vars() 13731 1727203885.26951: done getting variables 13731 1727203885.26995: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:96 Tuesday 24 September 2024 14:51:25 -0400 (0:00:00.037) 0:01:03.501 ***** 13731 1727203885.27024: entering _queue_task() for managed-node3/package 13731 1727203885.27270: worker is 1 (out of 1 available) 13731 1727203885.27286: exiting _queue_task() for managed-node3/package 13731 1727203885.27299: done queuing things up, now waiting for results queue to drain 13731 1727203885.27301: waiting for pending results... 13731 1727203885.27484: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable 13731 1727203885.27588: in run() - task 028d2410-947f-82dc-c122-000000000e16 13731 1727203885.27599: variable 'ansible_search_path' from source: unknown 13731 1727203885.27603: variable 'ansible_search_path' from source: unknown 13731 1727203885.27630: calling self._execute() 13731 1727203885.27708: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203885.27735: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203885.27744: variable 'omit' from source: magic vars 13731 1727203885.28267: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.28271: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203885.28380: variable 'network_state' from source: role '' defaults 13731 1727203885.28384: Evaluated conditional (network_state != {}): False 13731 1727203885.28386: when evaluation is False, skipping this task 13731 1727203885.28388: _execute() done 13731 1727203885.28390: dumping result to json 13731 1727203885.28392: done dumping result, returning 13731 1727203885.28398: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Install python3-libnmstate when using network_state variable [028d2410-947f-82dc-c122-000000000e16] 13731 1727203885.28409: sending task result for task 028d2410-947f-82dc-c122-000000000e16 skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203885.28574: no more pending results, returning what we have 13731 1727203885.28580: results queue empty 13731 1727203885.28581: checking for any_errors_fatal 13731 1727203885.28588: done checking for any_errors_fatal 13731 1727203885.28589: checking for max_fail_percentage 13731 1727203885.28591: done checking for max_fail_percentage 13731 1727203885.28592: checking to see if all hosts have failed and the running result is not ok 13731 1727203885.28592: done checking to see if all hosts have failed 13731 1727203885.28593: getting the remaining hosts for this loop 13731 1727203885.28594: done getting the remaining hosts for this loop 13731 1727203885.28598: getting the next task for host managed-node3 13731 1727203885.28607: done getting next task for host managed-node3 13731 1727203885.28610: ^ task is: TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13731 1727203885.28616: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203885.28639: getting variables 13731 1727203885.28641: in VariableManager get_vars() 13731 1727203885.28689: Calling all_inventory to load vars for managed-node3 13731 1727203885.28692: Calling groups_inventory to load vars for managed-node3 13731 1727203885.28694: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203885.28706: Calling all_plugins_play to load vars for managed-node3 13731 1727203885.28709: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203885.28712: Calling groups_plugins_play to load vars for managed-node3 13731 1727203885.29232: done sending task result for task 028d2410-947f-82dc-c122-000000000e16 13731 1727203885.29236: WORKER PROCESS EXITING 13731 1727203885.29940: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203885.31287: done with get_vars() 13731 1727203885.31313: done getting variables 13731 1727203885.31381: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:109 Tuesday 24 September 2024 14:51:25 -0400 (0:00:00.043) 0:01:03.545 ***** 13731 1727203885.31419: entering _queue_task() for managed-node3/service 13731 1727203885.31746: worker is 1 (out of 1 available) 13731 1727203885.31759: exiting _queue_task() for managed-node3/service 13731 1727203885.31771: done queuing things up, now waiting for results queue to drain 13731 1727203885.31772: waiting for pending results... 13731 1727203885.32083: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces 13731 1727203885.32258: in run() - task 028d2410-947f-82dc-c122-000000000e17 13731 1727203885.32285: variable 'ansible_search_path' from source: unknown 13731 1727203885.32289: variable 'ansible_search_path' from source: unknown 13731 1727203885.32320: calling self._execute() 13731 1727203885.32414: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203885.32421: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203885.32429: variable 'omit' from source: magic vars 13731 1727203885.32821: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.32825: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203885.33024: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203885.33163: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203885.35523: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203885.35605: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203885.35655: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203885.35696: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203885.35729: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203885.35848: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.35855: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.35888: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.35956: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.35959: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.36007: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.36034: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.36280: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.36284: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.36286: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.36289: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.36291: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.36294: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.36296: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.36298: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.36489: variable 'network_connections' from source: task vars 13731 1727203885.36510: variable 'port2_profile' from source: play vars 13731 1727203885.36592: variable 'port2_profile' from source: play vars 13731 1727203885.36610: variable 'port1_profile' from source: play vars 13731 1727203885.36682: variable 'port1_profile' from source: play vars 13731 1727203885.36695: variable 'controller_profile' from source: play vars 13731 1727203885.36763: variable 'controller_profile' from source: play vars 13731 1727203885.36835: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203885.37030: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203885.37080: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203885.37116: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203885.37147: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203885.37201: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203885.37225: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203885.37252: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.37392: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203885.37395: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203885.37650: variable 'network_connections' from source: task vars 13731 1727203885.37659: variable 'port2_profile' from source: play vars 13731 1727203885.37731: variable 'port2_profile' from source: play vars 13731 1727203885.37742: variable 'port1_profile' from source: play vars 13731 1727203885.37803: variable 'port1_profile' from source: play vars 13731 1727203885.37816: variable 'controller_profile' from source: play vars 13731 1727203885.37886: variable 'controller_profile' from source: play vars 13731 1727203885.37925: Evaluated conditional (__network_wireless_connections_defined or __network_team_connections_defined): False 13731 1727203885.37998: when evaluation is False, skipping this task 13731 1727203885.38001: _execute() done 13731 1727203885.38003: dumping result to json 13731 1727203885.38005: done dumping result, returning 13731 1727203885.38007: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Restart NetworkManager due to wireless or team interfaces [028d2410-947f-82dc-c122-000000000e17] 13731 1727203885.38009: sending task result for task 028d2410-947f-82dc-c122-000000000e17 skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wireless_connections_defined or __network_team_connections_defined", "skip_reason": "Conditional result was False" } 13731 1727203885.38262: no more pending results, returning what we have 13731 1727203885.38267: results queue empty 13731 1727203885.38268: checking for any_errors_fatal 13731 1727203885.38277: done checking for any_errors_fatal 13731 1727203885.38279: checking for max_fail_percentage 13731 1727203885.38281: done checking for max_fail_percentage 13731 1727203885.38282: checking to see if all hosts have failed and the running result is not ok 13731 1727203885.38283: done checking to see if all hosts have failed 13731 1727203885.38283: getting the remaining hosts for this loop 13731 1727203885.38285: done getting the remaining hosts for this loop 13731 1727203885.38290: getting the next task for host managed-node3 13731 1727203885.38300: done getting next task for host managed-node3 13731 1727203885.38304: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13731 1727203885.38311: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203885.38334: getting variables 13731 1727203885.38336: in VariableManager get_vars() 13731 1727203885.38504: Calling all_inventory to load vars for managed-node3 13731 1727203885.38508: Calling groups_inventory to load vars for managed-node3 13731 1727203885.38510: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203885.38521: Calling all_plugins_play to load vars for managed-node3 13731 1727203885.38524: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203885.38527: Calling groups_plugins_play to load vars for managed-node3 13731 1727203885.39116: done sending task result for task 028d2410-947f-82dc-c122-000000000e17 13731 1727203885.39119: WORKER PROCESS EXITING 13731 1727203885.40138: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203885.41700: done with get_vars() 13731 1727203885.41728: done getting variables 13731 1727203885.41793: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start NetworkManager] ***** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 Tuesday 24 September 2024 14:51:25 -0400 (0:00:00.104) 0:01:03.649 ***** 13731 1727203885.41829: entering _queue_task() for managed-node3/service 13731 1727203885.42391: worker is 1 (out of 1 available) 13731 1727203885.42401: exiting _queue_task() for managed-node3/service 13731 1727203885.42412: done queuing things up, now waiting for results queue to drain 13731 1727203885.42413: waiting for pending results... 13731 1727203885.42546: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager 13731 1727203885.42705: in run() - task 028d2410-947f-82dc-c122-000000000e18 13731 1727203885.42749: variable 'ansible_search_path' from source: unknown 13731 1727203885.42752: variable 'ansible_search_path' from source: unknown 13731 1727203885.42779: calling self._execute() 13731 1727203885.42968: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203885.42971: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203885.42976: variable 'omit' from source: magic vars 13731 1727203885.43291: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.43310: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203885.43482: variable 'network_provider' from source: set_fact 13731 1727203885.43493: variable 'network_state' from source: role '' defaults 13731 1727203885.43514: Evaluated conditional (network_provider == "nm" or network_state != {}): True 13731 1727203885.43524: variable 'omit' from source: magic vars 13731 1727203885.43601: variable 'omit' from source: magic vars 13731 1727203885.43641: variable 'network_service_name' from source: role '' defaults 13731 1727203885.43712: variable 'network_service_name' from source: role '' defaults 13731 1727203885.43828: variable '__network_provider_setup' from source: role '' defaults 13731 1727203885.43848: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203885.43912: variable '__network_service_name_default_nm' from source: role '' defaults 13731 1727203885.43925: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203885.44056: variable '__network_packages_default_nm' from source: role '' defaults 13731 1727203885.44220: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203885.46745: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203885.46828: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203885.46866: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203885.46909: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203885.46937: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203885.47029: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.47064: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.47101: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.47145: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.47209: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.47216: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.47245: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.47274: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.47325: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.47343: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.47587: variable '__network_packages_default_gobject_packages' from source: role '' defaults 13731 1727203885.47711: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.47753: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.47776: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.47862: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.47865: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.47935: variable 'ansible_python' from source: facts 13731 1727203885.47955: variable '__network_packages_default_wpa_supplicant' from source: role '' defaults 13731 1727203885.48045: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203885.48137: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203885.48277: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.48315: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.48480: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.48483: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.48485: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.48487: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203885.48497: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203885.48505: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.48547: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203885.48564: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203885.48709: variable 'network_connections' from source: task vars 13731 1727203885.48728: variable 'port2_profile' from source: play vars 13731 1727203885.48803: variable 'port2_profile' from source: play vars 13731 1727203885.48819: variable 'port1_profile' from source: play vars 13731 1727203885.48899: variable 'port1_profile' from source: play vars 13731 1727203885.48915: variable 'controller_profile' from source: play vars 13731 1727203885.48996: variable 'controller_profile' from source: play vars 13731 1727203885.49158: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203885.49324: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203885.49385: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203885.49429: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203885.49485: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203885.49555: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203885.49598: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203885.49682: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203885.49686: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203885.49736: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203885.50057: variable 'network_connections' from source: task vars 13731 1727203885.50069: variable 'port2_profile' from source: play vars 13731 1727203885.50160: variable 'port2_profile' from source: play vars 13731 1727203885.50181: variable 'port1_profile' from source: play vars 13731 1727203885.50266: variable 'port1_profile' from source: play vars 13731 1727203885.50345: variable 'controller_profile' from source: play vars 13731 1727203885.50369: variable 'controller_profile' from source: play vars 13731 1727203885.50411: variable '__network_packages_default_wireless' from source: role '' defaults 13731 1727203885.50563: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203885.50777: variable 'network_connections' from source: task vars 13731 1727203885.50795: variable 'port2_profile' from source: play vars 13731 1727203885.50858: variable 'port2_profile' from source: play vars 13731 1727203885.50895: variable 'port1_profile' from source: play vars 13731 1727203885.50944: variable 'port1_profile' from source: play vars 13731 1727203885.50956: variable 'controller_profile' from source: play vars 13731 1727203885.51029: variable 'controller_profile' from source: play vars 13731 1727203885.51113: variable '__network_packages_default_team' from source: role '' defaults 13731 1727203885.51143: variable '__network_team_connections_defined' from source: role '' defaults 13731 1727203885.51450: variable 'network_connections' from source: task vars 13731 1727203885.51463: variable 'port2_profile' from source: play vars 13731 1727203885.51546: variable 'port2_profile' from source: play vars 13731 1727203885.51561: variable 'port1_profile' from source: play vars 13731 1727203885.51627: variable 'port1_profile' from source: play vars 13731 1727203885.51639: variable 'controller_profile' from source: play vars 13731 1727203885.51712: variable 'controller_profile' from source: play vars 13731 1727203885.51872: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203885.51877: variable '__network_service_name_default_initscripts' from source: role '' defaults 13731 1727203885.51880: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203885.51898: variable '__network_packages_default_initscripts' from source: role '' defaults 13731 1727203885.52095: variable '__network_packages_default_initscripts_bridge' from source: role '' defaults 13731 1727203885.52580: variable 'network_connections' from source: task vars 13731 1727203885.52591: variable 'port2_profile' from source: play vars 13731 1727203885.52658: variable 'port2_profile' from source: play vars 13731 1727203885.52670: variable 'port1_profile' from source: play vars 13731 1727203885.52730: variable 'port1_profile' from source: play vars 13731 1727203885.52742: variable 'controller_profile' from source: play vars 13731 1727203885.52868: variable 'controller_profile' from source: play vars 13731 1727203885.52871: variable 'ansible_distribution' from source: facts 13731 1727203885.52873: variable '__network_rh_distros' from source: role '' defaults 13731 1727203885.52875: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.52880: variable '__network_packages_default_initscripts_network_scripts' from source: role '' defaults 13731 1727203885.53032: variable 'ansible_distribution' from source: facts 13731 1727203885.53042: variable '__network_rh_distros' from source: role '' defaults 13731 1727203885.53053: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.53073: variable '__network_packages_default_initscripts_dhcp_client' from source: role '' defaults 13731 1727203885.53269: variable 'ansible_distribution' from source: facts 13731 1727203885.53281: variable '__network_rh_distros' from source: role '' defaults 13731 1727203885.53291: variable 'ansible_distribution_major_version' from source: facts 13731 1727203885.53336: variable 'network_provider' from source: set_fact 13731 1727203885.53414: variable 'omit' from source: magic vars 13731 1727203885.53417: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203885.53439: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203885.53464: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203885.53489: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203885.53505: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203885.53547: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203885.53558: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203885.53568: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203885.53743: Set connection var ansible_pipelining to False 13731 1727203885.53746: Set connection var ansible_shell_type to sh 13731 1727203885.53748: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203885.53750: Set connection var ansible_connection to ssh 13731 1727203885.53752: Set connection var ansible_shell_executable to /bin/sh 13731 1727203885.53754: Set connection var ansible_timeout to 10 13731 1727203885.53771: variable 'ansible_shell_executable' from source: unknown 13731 1727203885.53781: variable 'ansible_connection' from source: unknown 13731 1727203885.53787: variable 'ansible_module_compression' from source: unknown 13731 1727203885.53793: variable 'ansible_shell_type' from source: unknown 13731 1727203885.53798: variable 'ansible_shell_executable' from source: unknown 13731 1727203885.53802: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203885.53808: variable 'ansible_pipelining' from source: unknown 13731 1727203885.53813: variable 'ansible_timeout' from source: unknown 13731 1727203885.53851: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203885.53936: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203885.53962: variable 'omit' from source: magic vars 13731 1727203885.53974: starting attempt loop 13731 1727203885.53982: running the handler 13731 1727203885.54177: variable 'ansible_facts' from source: unknown 13731 1727203885.54827: _low_level_execute_command(): starting 13731 1727203885.54847: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203885.55553: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203885.55569: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203885.55617: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203885.55633: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203885.55726: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203885.55741: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203885.55756: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203885.55835: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203885.57735: stdout chunk (state=3): >>>/root <<< 13731 1727203885.57893: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203885.57897: stdout chunk (state=3): >>><<< 13731 1727203885.57899: stderr chunk (state=3): >>><<< 13731 1727203885.57917: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203885.57935: _low_level_execute_command(): starting 13731 1727203885.58013: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036 `" && echo ansible-tmp-1727203885.5792377-18244-60049227036036="` echo /root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036 `" ) && sleep 0' 13731 1727203885.58542: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203885.58556: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203885.58570: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203885.58589: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203885.58605: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203885.58695: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203885.58711: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203885.58724: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203885.58743: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203885.58810: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203885.60691: stdout chunk (state=3): >>>ansible-tmp-1727203885.5792377-18244-60049227036036=/root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036 <<< 13731 1727203885.61059: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203885.61070: stdout chunk (state=3): >>><<< 13731 1727203885.61089: stderr chunk (state=3): >>><<< 13731 1727203885.61112: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203885.5792377-18244-60049227036036=/root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203885.61202: variable 'ansible_module_compression' from source: unknown 13731 1727203885.61217: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.systemd-ZIP_DEFLATED 13731 1727203885.61281: variable 'ansible_facts' from source: unknown 13731 1727203885.61511: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/AnsiballZ_systemd.py 13731 1727203885.61659: Sending initial data 13731 1727203885.61662: Sent initial data (155 bytes) 13731 1727203885.62398: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203885.62454: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203885.62465: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203885.62489: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203885.62544: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203885.64356: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203885.64402: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203885.64449: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp7echo6dl /root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/AnsiballZ_systemd.py <<< 13731 1727203885.64452: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/AnsiballZ_systemd.py" <<< 13731 1727203885.64488: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp7echo6dl" to remote "/root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/AnsiballZ_systemd.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/AnsiballZ_systemd.py" <<< 13731 1727203885.66185: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203885.66249: stderr chunk (state=3): >>><<< 13731 1727203885.66251: stdout chunk (state=3): >>><<< 13731 1727203885.66261: done transferring module to remote 13731 1727203885.66272: _low_level_execute_command(): starting 13731 1727203885.66277: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/ /root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/AnsiballZ_systemd.py && sleep 0' 13731 1727203885.66739: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203885.66780: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203885.66783: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203885.66786: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203885.66788: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203885.66833: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203885.66836: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203885.66881: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203885.68656: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203885.68660: stdout chunk (state=3): >>><<< 13731 1727203885.68702: stderr chunk (state=3): >>><<< 13731 1727203885.68718: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203885.68721: _low_level_execute_command(): starting 13731 1727203885.68745: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/AnsiballZ_systemd.py && sleep 0' 13731 1727203885.69280: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203885.69296: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203885.69312: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203885.69330: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203885.69348: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203885.69388: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203885.69472: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203885.69498: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203885.69569: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203885.98597: stdout chunk (state=3): >>> {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "RestartMode": "normal", "NotifyAccess": "none", "RestartUSec": "100ms", "RestartSteps": "0", "RestartMaxDelayUSec": "infinity", "RestartUSecNext": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "704", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "FileDescriptorStorePreserve": "restart", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainStartTimestampMonotonic": "28990148", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainHandoffTimestampMonotonic": "29005881", "ExecMainPID": "704", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "2938", "MemoryCurrent": "10469376", "MemoryPeak": "13586432", "MemorySwapCurrent": "0", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryAvailable": "3315601408", "EffectiveMemoryMax": "3702865920", "EffectiveMemoryHigh": "3702865920", "CPUUsageNSec": "1003294000", "TasksCurrent": "4", "EffectiveTasksMax": "22362", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultStartupMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "StartupMemoryLow": "0", "MemoryHigh": "infinity", "StartupMemoryHigh": "infinity", "MemoryMax": "infinity", "StartupMemoryMax": "infinity", "MemorySwapMax": "infinity", "StartupMemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22362", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "MemoryPressureWatch": "auto", "MemoryPressureThresholdUSec": "200ms", "CoredumpReceive": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "RootEphemeral": "no", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "SetLoginEnvironment": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "MemoryKSM": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "system.slice dbus.socket sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "cloud-init.service shutdown.target NetworkManager-wait-online.service multi-user.target network.target", "After": "dbus-b<<< 13731 1727203885.98729: stdout chunk (state=3): >>>roker.service systemd-journald.socket network-pre.target basic.target cloud-init-local.service dbus.socket system.slice sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:49:45 EDT", "StateChangeTimestampMonotonic": "362725592", "InactiveExitTimestamp": "Tue 2024-09-24 14:44:11 EDT", "InactiveExitTimestampMonotonic": "28990654", "ActiveEnterTimestamp": "Tue 2024-09-24 14:44:12 EDT", "ActiveEnterTimestampMonotonic": "29769382", "ActiveExitTimestampMonotonic": "0", "InactiveEnterTimestampMonotonic": "0", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "SurviveFinalKillSignal": "no", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ConditionTimestampMonotonic": "28989295", "AssertTimestamp": "Tue 2024-09-24 14:44:11 EDT", "AssertTimestampMonotonic": "28989297", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "70a845f8a1964db89963090ed497f47f", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 13731 1727203886.00356: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203886.00380: stderr chunk (state=3): >>>Shared connection to 10.31.47.22 closed. <<< 13731 1727203886.00438: stderr chunk (state=3): >>><<< 13731 1727203886.00441: stdout chunk (state=3): >>><<< 13731 1727203886.00466: _low_level_execute_command() done: rc=0, stdout= {"name": "NetworkManager", "changed": false, "status": {"Type": "dbus", "ExitType": "main", "Restart": "on-failure", "RestartMode": "normal", "NotifyAccess": "none", "RestartUSec": "100ms", "RestartSteps": "0", "RestartMaxDelayUSec": "infinity", "RestartUSecNext": "100ms", "TimeoutStartUSec": "10min", "TimeoutStopUSec": "1min 30s", "TimeoutAbortUSec": "1min 30s", "TimeoutStartFailureMode": "terminate", "TimeoutStopFailureMode": "terminate", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "WatchdogUSec": "0", "WatchdogTimestampMonotonic": "0", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "704", "ControlPID": "0", "BusName": "org.freedesktop.NetworkManager", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "FileDescriptorStorePreserve": "restart", "StatusErrno": "0", "Result": "success", "ReloadResult": "success", "CleanResult": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "OOMPolicy": "stop", "ReloadSignal": "1", "ExecMainStartTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainStartTimestampMonotonic": "28990148", "ExecMainExitTimestampMonotonic": "0", "ExecMainHandoffTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ExecMainHandoffTimestampMonotonic": "29005881", "ExecMainPID": "704", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/sbin/NetworkManager ; argv[]=/usr/sbin/NetworkManager --no-daemon ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReload": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecReloadEx": "{ path=/usr/bin/busctl ; argv[]=/usr/bin/busctl call org.freedesktop.NetworkManager /org/freedesktop/NetworkManager org.freedesktop.NetworkManager Reload u 0 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/NetworkManager.service", "ControlGroupId": "2938", "MemoryCurrent": "10469376", "MemoryPeak": "13586432", "MemorySwapCurrent": "0", "MemorySwapPeak": "0", "MemoryZSwapCurrent": "0", "MemoryAvailable": "3315601408", "EffectiveMemoryMax": "3702865920", "EffectiveMemoryHigh": "3702865920", "CPUUsageNSec": "1003294000", "TasksCurrent": "4", "EffectiveTasksMax": "22362", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IOReadBytes": "[not set]", "IOReadOperations": "[not set]", "IOWriteBytes": "[not set]", "IOWriteOperations": "[not set]", "Delegate": "no", "CPUAccounting": "yes", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultStartupMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "StartupMemoryLow": "0", "MemoryHigh": "infinity", "StartupMemoryHigh": "infinity", "MemoryMax": "infinity", "StartupMemoryMax": "infinity", "MemorySwapMax": "infinity", "StartupMemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "StartupMemoryZSwapMax": "infinity", "MemoryZSwapWriteback": "yes", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22362", "IPAccounting": "no", "ManagedOOMSwap": "auto", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "MemoryPressureWatch": "auto", "MemoryPressureThresholdUSec": "200ms", "CoredumpReceive": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "65536", "LimitNOFILESoft": "65536", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "13976", "LimitNPROCSoft": "13976", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "13976", "LimitSIGPENDINGSoft": "13976", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "RootEphemeral": "no", "OOMScoreAdjust": "0", "CoredumpFilter": "0x33", "Nice": "0", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_dac_override cap_kill cap_setgid cap_setuid cap_net_bind_service cap_net_admin cap_net_raw cap_sys_module cap_sys_chroot cap_audit_write", "DynamicUser": "no", "SetLoginEnvironment": "no", "RemoveIPC": "no", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectClock": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectKernelLogs": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "PrivateIPC": "no", "ProtectHome": "read-only", "ProtectSystem": "yes", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "2147483646", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "TimeoutCleanUSec": "infinity", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "ProtectProc": "default", "ProcSubset": "all", "ProtectHostname": "no", "MemoryKSM": "no", "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent", "KillMode": "process", "KillSignal": "15", "RestartKillSignal": "15", "FinalKillSignal": "9", "SendSIGKILL": "yes", "SendSIGHUP": "no", "WatchdogSignal": "6", "Id": "NetworkManager.service", "Names": "NetworkManager.service", "Requires": "system.slice dbus.socket sysinit.target", "Wants": "network.target", "BindsTo": "dbus-broker.service", "RequiredBy": "NetworkManager-wait-online.service", "WantedBy": "multi-user.target", "Conflicts": "shutdown.target", "Before": "cloud-init.service shutdown.target NetworkManager-wait-online.service multi-user.target network.target", "After": "dbus-broker.service systemd-journald.socket network-pre.target basic.target cloud-init-local.service dbus.socket system.slice sysinit.target", "Documentation": "\"man:NetworkManager(8)\"", "Description": "Network Manager", "AccessSELinuxContext": "system_u:object_r:NetworkManager_unit_file_t:s0", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/NetworkManager.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Tue 2024-09-24 14:49:45 EDT", "StateChangeTimestampMonotonic": "362725592", "InactiveExitTimestamp": "Tue 2024-09-24 14:44:11 EDT", "InactiveExitTimestampMonotonic": "28990654", "ActiveEnterTimestamp": "Tue 2024-09-24 14:44:12 EDT", "ActiveEnterTimestampMonotonic": "29769382", "ActiveExitTimestampMonotonic": "0", "InactiveEnterTimestampMonotonic": "0", "CanStart": "yes", "CanStop": "yes", "CanReload": "yes", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "SurviveFinalKillSignal": "no", "OnSuccessJobMode": "fail", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Tue 2024-09-24 14:44:11 EDT", "ConditionTimestampMonotonic": "28989295", "AssertTimestamp": "Tue 2024-09-24 14:44:11 EDT", "AssertTimestampMonotonic": "28989297", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "70a845f8a1964db89963090ed497f47f", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "NetworkManager", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203886.00647: done with _execute_module (ansible.legacy.systemd, {'name': 'NetworkManager', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': True, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203886.00656: _low_level_execute_command(): starting 13731 1727203886.00663: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203885.5792377-18244-60049227036036/ > /dev/null 2>&1 && sleep 0' 13731 1727203886.01259: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203886.01266: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203886.01281: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203886.01294: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203886.01304: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203886.01329: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203886.01333: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203886.01352: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203886.01445: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203886.01503: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203886.02082: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203886.03385: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203886.03430: stderr chunk (state=3): >>><<< 13731 1727203886.03436: stdout chunk (state=3): >>><<< 13731 1727203886.03457: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203886.03465: handler run complete 13731 1727203886.03528: attempt loop complete, returning result 13731 1727203886.03532: _execute() done 13731 1727203886.03534: dumping result to json 13731 1727203886.03551: done dumping result, returning 13731 1727203886.03563: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start NetworkManager [028d2410-947f-82dc-c122-000000000e18] 13731 1727203886.03566: sending task result for task 028d2410-947f-82dc-c122-000000000e18 ok: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203886.04340: no more pending results, returning what we have 13731 1727203886.04343: results queue empty 13731 1727203886.04344: checking for any_errors_fatal 13731 1727203886.04347: done checking for any_errors_fatal 13731 1727203886.04348: checking for max_fail_percentage 13731 1727203886.04349: done checking for max_fail_percentage 13731 1727203886.04350: checking to see if all hosts have failed and the running result is not ok 13731 1727203886.04351: done checking to see if all hosts have failed 13731 1727203886.04352: getting the remaining hosts for this loop 13731 1727203886.04353: done getting the remaining hosts for this loop 13731 1727203886.04356: getting the next task for host managed-node3 13731 1727203886.04365: done getting next task for host managed-node3 13731 1727203886.04369: ^ task is: TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13731 1727203886.04378: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203886.04389: getting variables 13731 1727203886.04391: in VariableManager get_vars() 13731 1727203886.04429: Calling all_inventory to load vars for managed-node3 13731 1727203886.04432: Calling groups_inventory to load vars for managed-node3 13731 1727203886.04434: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203886.04443: Calling all_plugins_play to load vars for managed-node3 13731 1727203886.04446: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203886.04448: Calling groups_plugins_play to load vars for managed-node3 13731 1727203886.05026: done sending task result for task 028d2410-947f-82dc-c122-000000000e18 13731 1727203886.05030: WORKER PROCESS EXITING 13731 1727203886.06519: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203886.08206: done with get_vars() 13731 1727203886.08230: done getting variables 13731 1727203886.08303: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable and start wpa_supplicant] ***** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:133 Tuesday 24 September 2024 14:51:26 -0400 (0:00:00.665) 0:01:04.314 ***** 13731 1727203886.08347: entering _queue_task() for managed-node3/service 13731 1727203886.08731: worker is 1 (out of 1 available) 13731 1727203886.08747: exiting _queue_task() for managed-node3/service 13731 1727203886.08759: done queuing things up, now waiting for results queue to drain 13731 1727203886.08764: waiting for pending results... 13731 1727203886.08979: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant 13731 1727203886.09229: in run() - task 028d2410-947f-82dc-c122-000000000e19 13731 1727203886.09266: variable 'ansible_search_path' from source: unknown 13731 1727203886.09359: variable 'ansible_search_path' from source: unknown 13731 1727203886.09366: calling self._execute() 13731 1727203886.09439: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203886.09451: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203886.09484: variable 'omit' from source: magic vars 13731 1727203886.09918: variable 'ansible_distribution_major_version' from source: facts 13731 1727203886.09936: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203886.10067: variable 'network_provider' from source: set_fact 13731 1727203886.10081: Evaluated conditional (network_provider == "nm"): True 13731 1727203886.10183: variable '__network_wpa_supplicant_required' from source: role '' defaults 13731 1727203886.10288: variable '__network_ieee802_1x_connections_defined' from source: role '' defaults 13731 1727203886.10480: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203886.12500: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203886.12544: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203886.12572: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203886.12599: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203886.12623: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203886.12806: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203886.12829: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203886.12847: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203886.12874: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203886.12886: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203886.12918: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203886.12936: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203886.12955: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203886.12982: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203886.12993: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203886.13020: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203886.13037: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203886.13057: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203886.13083: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203886.13093: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203886.13192: variable 'network_connections' from source: task vars 13731 1727203886.13202: variable 'port2_profile' from source: play vars 13731 1727203886.13249: variable 'port2_profile' from source: play vars 13731 1727203886.13266: variable 'port1_profile' from source: play vars 13731 1727203886.13307: variable 'port1_profile' from source: play vars 13731 1727203886.13313: variable 'controller_profile' from source: play vars 13731 1727203886.13354: variable 'controller_profile' from source: play vars 13731 1727203886.13406: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 13731 1727203886.13519: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 13731 1727203886.13545: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 13731 1727203886.13570: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 13731 1727203886.13598: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 13731 1727203886.13627: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 13731 1727203886.13642: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 13731 1727203886.13658: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203886.13678: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 13731 1727203886.13718: variable '__network_wireless_connections_defined' from source: role '' defaults 13731 1727203886.14080: variable 'network_connections' from source: task vars 13731 1727203886.14083: variable 'port2_profile' from source: play vars 13731 1727203886.14086: variable 'port2_profile' from source: play vars 13731 1727203886.14088: variable 'port1_profile' from source: play vars 13731 1727203886.14090: variable 'port1_profile' from source: play vars 13731 1727203886.14100: variable 'controller_profile' from source: play vars 13731 1727203886.14153: variable 'controller_profile' from source: play vars 13731 1727203886.14195: Evaluated conditional (__network_wpa_supplicant_required): False 13731 1727203886.14204: when evaluation is False, skipping this task 13731 1727203886.14212: _execute() done 13731 1727203886.14230: dumping result to json 13731 1727203886.14238: done dumping result, returning 13731 1727203886.14249: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable and start wpa_supplicant [028d2410-947f-82dc-c122-000000000e19] 13731 1727203886.14262: sending task result for task 028d2410-947f-82dc-c122-000000000e19 skipping: [managed-node3] => { "changed": false, "false_condition": "__network_wpa_supplicant_required", "skip_reason": "Conditional result was False" } 13731 1727203886.14417: no more pending results, returning what we have 13731 1727203886.14422: results queue empty 13731 1727203886.14423: checking for any_errors_fatal 13731 1727203886.14448: done checking for any_errors_fatal 13731 1727203886.14449: checking for max_fail_percentage 13731 1727203886.14451: done checking for max_fail_percentage 13731 1727203886.14452: checking to see if all hosts have failed and the running result is not ok 13731 1727203886.14452: done checking to see if all hosts have failed 13731 1727203886.14453: getting the remaining hosts for this loop 13731 1727203886.14455: done getting the remaining hosts for this loop 13731 1727203886.14459: getting the next task for host managed-node3 13731 1727203886.14471: done getting next task for host managed-node3 13731 1727203886.14478: ^ task is: TASK: fedora.linux_system_roles.network : Enable network service 13731 1727203886.14483: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203886.14506: getting variables 13731 1727203886.14508: in VariableManager get_vars() 13731 1727203886.14554: Calling all_inventory to load vars for managed-node3 13731 1727203886.14557: Calling groups_inventory to load vars for managed-node3 13731 1727203886.14560: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203886.14573: Calling all_plugins_play to load vars for managed-node3 13731 1727203886.14795: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203886.14802: done sending task result for task 028d2410-947f-82dc-c122-000000000e19 13731 1727203886.14805: WORKER PROCESS EXITING 13731 1727203886.14809: Calling groups_plugins_play to load vars for managed-node3 13731 1727203886.15785: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203886.16678: done with get_vars() 13731 1727203886.16694: done getting variables 13731 1727203886.16743: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Enable network service] ************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:142 Tuesday 24 September 2024 14:51:26 -0400 (0:00:00.084) 0:01:04.398 ***** 13731 1727203886.16770: entering _queue_task() for managed-node3/service 13731 1727203886.17030: worker is 1 (out of 1 available) 13731 1727203886.17045: exiting _queue_task() for managed-node3/service 13731 1727203886.17071: done queuing things up, now waiting for results queue to drain 13731 1727203886.17073: waiting for pending results... 13731 1727203886.17311: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable network service 13731 1727203886.17504: in run() - task 028d2410-947f-82dc-c122-000000000e1a 13731 1727203886.17525: variable 'ansible_search_path' from source: unknown 13731 1727203886.17533: variable 'ansible_search_path' from source: unknown 13731 1727203886.17580: calling self._execute() 13731 1727203886.17692: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203886.17711: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203886.17730: variable 'omit' from source: magic vars 13731 1727203886.18146: variable 'ansible_distribution_major_version' from source: facts 13731 1727203886.18170: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203886.18303: variable 'network_provider' from source: set_fact 13731 1727203886.18313: Evaluated conditional (network_provider == "initscripts"): False 13731 1727203886.18367: when evaluation is False, skipping this task 13731 1727203886.18374: _execute() done 13731 1727203886.18378: dumping result to json 13731 1727203886.18380: done dumping result, returning 13731 1727203886.18383: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Enable network service [028d2410-947f-82dc-c122-000000000e1a] 13731 1727203886.18386: sending task result for task 028d2410-947f-82dc-c122-000000000e1a skipping: [managed-node3] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false } 13731 1727203886.18532: no more pending results, returning what we have 13731 1727203886.18537: results queue empty 13731 1727203886.18537: checking for any_errors_fatal 13731 1727203886.18546: done checking for any_errors_fatal 13731 1727203886.18547: checking for max_fail_percentage 13731 1727203886.18549: done checking for max_fail_percentage 13731 1727203886.18550: checking to see if all hosts have failed and the running result is not ok 13731 1727203886.18550: done checking to see if all hosts have failed 13731 1727203886.18551: getting the remaining hosts for this loop 13731 1727203886.18553: done getting the remaining hosts for this loop 13731 1727203886.18556: getting the next task for host managed-node3 13731 1727203886.18567: done getting next task for host managed-node3 13731 1727203886.18570: ^ task is: TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13731 1727203886.18583: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=19, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203886.18608: getting variables 13731 1727203886.18610: in VariableManager get_vars() 13731 1727203886.18655: Calling all_inventory to load vars for managed-node3 13731 1727203886.18658: Calling groups_inventory to load vars for managed-node3 13731 1727203886.18663: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203886.18674: Calling all_plugins_play to load vars for managed-node3 13731 1727203886.18679: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203886.18685: done sending task result for task 028d2410-947f-82dc-c122-000000000e1a 13731 1727203886.18690: WORKER PROCESS EXITING 13731 1727203886.18697: Calling groups_plugins_play to load vars for managed-node3 13731 1727203886.19513: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203886.20459: done with get_vars() 13731 1727203886.20487: done getting variables 13731 1727203886.20541: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Ensure initscripts network file dependency is present] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:150 Tuesday 24 September 2024 14:51:26 -0400 (0:00:00.038) 0:01:04.436 ***** 13731 1727203886.20579: entering _queue_task() for managed-node3/copy 13731 1727203886.21012: worker is 1 (out of 1 available) 13731 1727203886.21022: exiting _queue_task() for managed-node3/copy 13731 1727203886.21036: done queuing things up, now waiting for results queue to drain 13731 1727203886.21037: waiting for pending results... 13731 1727203886.21249: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present 13731 1727203886.21401: in run() - task 028d2410-947f-82dc-c122-000000000e1b 13731 1727203886.21481: variable 'ansible_search_path' from source: unknown 13731 1727203886.21485: variable 'ansible_search_path' from source: unknown 13731 1727203886.21488: calling self._execute() 13731 1727203886.21565: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203886.21580: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203886.21596: variable 'omit' from source: magic vars 13731 1727203886.21981: variable 'ansible_distribution_major_version' from source: facts 13731 1727203886.21998: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203886.22115: variable 'network_provider' from source: set_fact 13731 1727203886.22126: Evaluated conditional (network_provider == "initscripts"): False 13731 1727203886.22134: when evaluation is False, skipping this task 13731 1727203886.22141: _execute() done 13731 1727203886.22280: dumping result to json 13731 1727203886.22283: done dumping result, returning 13731 1727203886.22287: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Ensure initscripts network file dependency is present [028d2410-947f-82dc-c122-000000000e1b] 13731 1727203886.22289: sending task result for task 028d2410-947f-82dc-c122-000000000e1b 13731 1727203886.22359: done sending task result for task 028d2410-947f-82dc-c122-000000000e1b 13731 1727203886.22365: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_provider == \"initscripts\"", "skip_reason": "Conditional result was False" } 13731 1727203886.22420: no more pending results, returning what we have 13731 1727203886.22424: results queue empty 13731 1727203886.22425: checking for any_errors_fatal 13731 1727203886.22434: done checking for any_errors_fatal 13731 1727203886.22434: checking for max_fail_percentage 13731 1727203886.22436: done checking for max_fail_percentage 13731 1727203886.22437: checking to see if all hosts have failed and the running result is not ok 13731 1727203886.22437: done checking to see if all hosts have failed 13731 1727203886.22438: getting the remaining hosts for this loop 13731 1727203886.22440: done getting the remaining hosts for this loop 13731 1727203886.22443: getting the next task for host managed-node3 13731 1727203886.22451: done getting next task for host managed-node3 13731 1727203886.22455: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13731 1727203886.22464: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=20, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203886.22493: getting variables 13731 1727203886.22496: in VariableManager get_vars() 13731 1727203886.22551: Calling all_inventory to load vars for managed-node3 13731 1727203886.22554: Calling groups_inventory to load vars for managed-node3 13731 1727203886.22556: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203886.22569: Calling all_plugins_play to load vars for managed-node3 13731 1727203886.22571: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203886.22574: Calling groups_plugins_play to load vars for managed-node3 13731 1727203886.23932: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203886.25080: done with get_vars() 13731 1727203886.25104: done getting variables TASK [fedora.linux_system_roles.network : Configure networking connection profiles] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 Tuesday 24 September 2024 14:51:26 -0400 (0:00:00.046) 0:01:04.482 ***** 13731 1727203886.25195: entering _queue_task() for managed-node3/fedora.linux_system_roles.network_connections 13731 1727203886.25528: worker is 1 (out of 1 available) 13731 1727203886.25541: exiting _queue_task() for managed-node3/fedora.linux_system_roles.network_connections 13731 1727203886.25557: done queuing things up, now waiting for results queue to drain 13731 1727203886.25558: waiting for pending results... 13731 1727203886.26283: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking connection profiles 13731 1727203886.26357: in run() - task 028d2410-947f-82dc-c122-000000000e1c 13731 1727203886.26386: variable 'ansible_search_path' from source: unknown 13731 1727203886.26395: variable 'ansible_search_path' from source: unknown 13731 1727203886.26434: calling self._execute() 13731 1727203886.26547: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203886.26558: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203886.26577: variable 'omit' from source: magic vars 13731 1727203886.26963: variable 'ansible_distribution_major_version' from source: facts 13731 1727203886.26982: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203886.26992: variable 'omit' from source: magic vars 13731 1727203886.27070: variable 'omit' from source: magic vars 13731 1727203886.27235: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 13731 1727203886.29482: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 13731 1727203886.29634: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 13731 1727203886.29637: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 13731 1727203886.29639: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 13731 1727203886.29666: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 13731 1727203886.29759: variable 'network_provider' from source: set_fact 13731 1727203886.29904: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 13731 1727203886.29937: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 13731 1727203886.29978: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 13731 1727203886.30022: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 13731 1727203886.30041: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 13731 1727203886.30127: variable 'omit' from source: magic vars 13731 1727203886.30249: variable 'omit' from source: magic vars 13731 1727203886.30358: variable 'network_connections' from source: task vars 13731 1727203886.30390: variable 'port2_profile' from source: play vars 13731 1727203886.30444: variable 'port2_profile' from source: play vars 13731 1727203886.30499: variable 'port1_profile' from source: play vars 13731 1727203886.30528: variable 'port1_profile' from source: play vars 13731 1727203886.30539: variable 'controller_profile' from source: play vars 13731 1727203886.30608: variable 'controller_profile' from source: play vars 13731 1727203886.30780: variable 'omit' from source: magic vars 13731 1727203886.30793: variable '__lsr_ansible_managed' from source: task vars 13731 1727203886.30859: variable '__lsr_ansible_managed' from source: task vars 13731 1727203886.31152: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 13731 1727203886.31306: Loaded config def from plugin (lookup/template) 13731 1727203886.31317: Loading LookupModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/template.py 13731 1727203886.31346: File lookup term: get_ansible_managed.j2 13731 1727203886.31353: variable 'ansible_search_path' from source: unknown 13731 1727203886.31368: evaluation_path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks 13731 1727203886.31388: search_path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/templates/get_ansible_managed.j2 /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/get_ansible_managed.j2 13731 1727203886.31409: variable 'ansible_search_path' from source: unknown 13731 1727203886.37917: variable 'ansible_managed' from source: unknown 13731 1727203886.38073: variable 'omit' from source: magic vars 13731 1727203886.38110: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203886.38143: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203886.38173: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203886.38270: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203886.38273: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203886.38278: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203886.38280: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203886.38282: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203886.38346: Set connection var ansible_pipelining to False 13731 1727203886.38355: Set connection var ansible_shell_type to sh 13731 1727203886.38366: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203886.38380: Set connection var ansible_connection to ssh 13731 1727203886.38388: Set connection var ansible_shell_executable to /bin/sh 13731 1727203886.38396: Set connection var ansible_timeout to 10 13731 1727203886.38418: variable 'ansible_shell_executable' from source: unknown 13731 1727203886.38425: variable 'ansible_connection' from source: unknown 13731 1727203886.38431: variable 'ansible_module_compression' from source: unknown 13731 1727203886.38437: variable 'ansible_shell_type' from source: unknown 13731 1727203886.38443: variable 'ansible_shell_executable' from source: unknown 13731 1727203886.38450: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203886.38457: variable 'ansible_pipelining' from source: unknown 13731 1727203886.38469: variable 'ansible_timeout' from source: unknown 13731 1727203886.38580: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203886.38631: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203886.38647: variable 'omit' from source: magic vars 13731 1727203886.38659: starting attempt loop 13731 1727203886.38667: running the handler 13731 1727203886.38687: _low_level_execute_command(): starting 13731 1727203886.38697: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203886.39411: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203886.39426: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203886.39444: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203886.39480: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203886.39587: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203886.39606: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203886.39687: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203886.41347: stdout chunk (state=3): >>>/root <<< 13731 1727203886.41489: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203886.41508: stdout chunk (state=3): >>><<< 13731 1727203886.41633: stderr chunk (state=3): >>><<< 13731 1727203886.41638: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203886.41641: _low_level_execute_command(): starting 13731 1727203886.41643: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040 `" && echo ansible-tmp-1727203886.4154372-18293-71022728186040="` echo /root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040 `" ) && sleep 0' 13731 1727203886.42211: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203886.42225: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203886.42241: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203886.42291: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203886.42365: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203886.42384: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203886.42407: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203886.42476: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203886.44339: stdout chunk (state=3): >>>ansible-tmp-1727203886.4154372-18293-71022728186040=/root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040 <<< 13731 1727203886.44502: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203886.44506: stdout chunk (state=3): >>><<< 13731 1727203886.44508: stderr chunk (state=3): >>><<< 13731 1727203886.44681: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203886.4154372-18293-71022728186040=/root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203886.44684: variable 'ansible_module_compression' from source: unknown 13731 1727203886.44686: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.network_connections-ZIP_DEFLATED 13731 1727203886.44689: variable 'ansible_facts' from source: unknown 13731 1727203886.44834: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/AnsiballZ_network_connections.py 13731 1727203886.45043: Sending initial data 13731 1727203886.45046: Sent initial data (167 bytes) 13731 1727203886.45645: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203886.45692: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203886.45712: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203886.45796: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203886.45815: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203886.45837: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203886.45988: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203886.47497: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203886.47549: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203886.47586: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmppmp7mip_ /root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/AnsiballZ_network_connections.py <<< 13731 1727203886.47590: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/AnsiballZ_network_connections.py" <<< 13731 1727203886.47625: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmppmp7mip_" to remote "/root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/AnsiballZ_network_connections.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/AnsiballZ_network_connections.py" <<< 13731 1727203886.48919: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203886.48923: stderr chunk (state=3): >>><<< 13731 1727203886.48925: stdout chunk (state=3): >>><<< 13731 1727203886.48948: done transferring module to remote 13731 1727203886.48959: _low_level_execute_command(): starting 13731 1727203886.48968: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/ /root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/AnsiballZ_network_connections.py && sleep 0' 13731 1727203886.49847: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203886.49851: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203886.49891: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203886.49967: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203886.50010: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203886.50054: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203886.51959: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203886.51965: stdout chunk (state=3): >>><<< 13731 1727203886.51967: stderr chunk (state=3): >>><<< 13731 1727203886.52040: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203886.52043: _low_level_execute_command(): starting 13731 1727203886.52046: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/AnsiballZ_network_connections.py && sleep 0' 13731 1727203886.52538: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203886.52546: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203886.52556: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203886.52569: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203886.52581: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203886.52589: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203886.52597: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203886.52610: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203886.52680: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203886.52688: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203886.52692: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203886.52694: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203886.52696: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203886.52698: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203886.52700: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203886.52702: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203886.52726: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203886.52742: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203886.52818: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203886.52823: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.05636: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.1/9eb2a8f6-720f-41da-85c9-7f6979308b82: error=unknown <<< 13731 1727203887.07640: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail <<< 13731 1727203887.07667: stdout chunk (state=3): >>>ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.0/76f9383b-e1a2-4403-937c-02353034d999: error=unknown <<< 13731 1727203887.09358: stdout chunk (state=3): >>>Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0/65e7c30c-f406-45a4-954f-9e2a9480d1df: error=unknown <<< 13731 1727203887.09620: stdout chunk (state=3): >>> <<< 13731 1727203887.09633: stdout chunk (state=3): >>>{"changed": true, "warnings": [], "stderr": "\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} <<< 13731 1727203887.11500: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203887.11522: stdout chunk (state=3): >>><<< 13731 1727203887.11525: stderr chunk (state=3): >>><<< 13731 1727203887.11680: _low_level_execute_command() done: rc=0, stdout=Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.1/9eb2a8f6-720f-41da-85c9-7f6979308b82: error=unknown Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0.0/76f9383b-e1a2-4403-937c-02353034d999: error=unknown Traceback (most recent call last): File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/connection.py", line 113, in _nm_profile_volatile_update2_call_back File "/tmp/ansible_fedora.linux_system_roles.network_connections_payload_ahwuv_ya/ansible_fedora.linux_system_roles.network_connections_payload.zip/ansible_collections/fedora/linux_system_roles/plugins/module_utils/network_lsr/nm/client.py", line 102, in fail ansible_collections.fedora.linux_system_roles.plugins.module_utils.network_lsr.nm.error.LsrNetworkNmError: Connection volatilize aborted on bond0/65e7c30c-f406-45a4-954f-9e2a9480d1df: error=unknown {"changed": true, "warnings": [], "stderr": "\n", "_invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}, "invocation": {"module_args": {"provider": "nm", "connections": [{"name": "bond0.1", "persistent_state": "absent", "state": "down"}, {"name": "bond0.0", "persistent_state": "absent", "state": "down"}, {"name": "bond0", "persistent_state": "absent", "state": "down"}], "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "ignore_errors": false, "force_state_change": false, "__debug_flags": ""}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203887.11690: done with _execute_module (fedora.linux_system_roles.network_connections, {'provider': 'nm', 'connections': [{'name': 'bond0.1', 'persistent_state': 'absent', 'state': 'down'}, {'name': 'bond0.0', 'persistent_state': 'absent', 'state': 'down'}, {'name': 'bond0', 'persistent_state': 'absent', 'state': 'down'}], '__header': '#\n# Ansible managed\n#\n# system_role:network\n', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'fedora.linux_system_roles.network_connections', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203887.11693: _low_level_execute_command(): starting 13731 1727203887.11880: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203886.4154372-18293-71022728186040/ > /dev/null 2>&1 && sleep 0' 13731 1727203887.12242: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203887.12259: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.12279: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203887.12297: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203887.12315: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203887.12328: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203887.12343: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.12366: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203887.12385: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203887.12401: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203887.12414: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.12428: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203887.12444: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203887.12524: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.12546: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.12609: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.14438: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.14501: stderr chunk (state=3): >>><<< 13731 1727203887.14510: stdout chunk (state=3): >>><<< 13731 1727203887.14529: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203887.14540: handler run complete 13731 1727203887.14586: attempt loop complete, returning result 13731 1727203887.14594: _execute() done 13731 1727203887.14601: dumping result to json 13731 1727203887.14611: done dumping result, returning 13731 1727203887.14624: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking connection profiles [028d2410-947f-82dc-c122-000000000e1c] 13731 1727203887.14631: sending task result for task 028d2410-947f-82dc-c122-000000000e1c changed: [managed-node3] => { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "name": "bond0.1", "persistent_state": "absent", "state": "down" }, { "name": "bond0.0", "persistent_state": "absent", "state": "down" }, { "name": "bond0", "persistent_state": "absent", "state": "down" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true } STDERR: 13731 1727203887.14888: no more pending results, returning what we have 13731 1727203887.14893: results queue empty 13731 1727203887.14893: checking for any_errors_fatal 13731 1727203887.14900: done checking for any_errors_fatal 13731 1727203887.14901: checking for max_fail_percentage 13731 1727203887.14903: done checking for max_fail_percentage 13731 1727203887.14903: checking to see if all hosts have failed and the running result is not ok 13731 1727203887.14904: done checking to see if all hosts have failed 13731 1727203887.14905: getting the remaining hosts for this loop 13731 1727203887.14907: done getting the remaining hosts for this loop 13731 1727203887.14910: getting the next task for host managed-node3 13731 1727203887.14918: done getting next task for host managed-node3 13731 1727203887.14922: ^ task is: TASK: fedora.linux_system_roles.network : Configure networking state 13731 1727203887.14927: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=21, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203887.14944: getting variables 13731 1727203887.14946: in VariableManager get_vars() 13731 1727203887.15202: Calling all_inventory to load vars for managed-node3 13731 1727203887.15206: Calling groups_inventory to load vars for managed-node3 13731 1727203887.15208: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203887.15303: Calling all_plugins_play to load vars for managed-node3 13731 1727203887.15307: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203887.15310: Calling groups_plugins_play to load vars for managed-node3 13731 1727203887.15913: done sending task result for task 028d2410-947f-82dc-c122-000000000e1c 13731 1727203887.15918: WORKER PROCESS EXITING 13731 1727203887.16868: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203887.18701: done with get_vars() 13731 1727203887.18723: done getting variables TASK [fedora.linux_system_roles.network : Configure networking state] ********** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:171 Tuesday 24 September 2024 14:51:27 -0400 (0:00:00.936) 0:01:05.419 ***** 13731 1727203887.18822: entering _queue_task() for managed-node3/fedora.linux_system_roles.network_state 13731 1727203887.19185: worker is 1 (out of 1 available) 13731 1727203887.19283: exiting _queue_task() for managed-node3/fedora.linux_system_roles.network_state 13731 1727203887.19294: done queuing things up, now waiting for results queue to drain 13731 1727203887.19295: waiting for pending results... 13731 1727203887.19524: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking state 13731 1727203887.19717: in run() - task 028d2410-947f-82dc-c122-000000000e1d 13731 1727203887.19739: variable 'ansible_search_path' from source: unknown 13731 1727203887.19753: variable 'ansible_search_path' from source: unknown 13731 1727203887.19796: calling self._execute() 13731 1727203887.19902: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.19913: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.19927: variable 'omit' from source: magic vars 13731 1727203887.20334: variable 'ansible_distribution_major_version' from source: facts 13731 1727203887.20349: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203887.20487: variable 'network_state' from source: role '' defaults 13731 1727203887.20504: Evaluated conditional (network_state != {}): False 13731 1727203887.20517: when evaluation is False, skipping this task 13731 1727203887.20524: _execute() done 13731 1727203887.20580: dumping result to json 13731 1727203887.20583: done dumping result, returning 13731 1727203887.20586: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Configure networking state [028d2410-947f-82dc-c122-000000000e1d] 13731 1727203887.20589: sending task result for task 028d2410-947f-82dc-c122-000000000e1d 13731 1727203887.20731: done sending task result for task 028d2410-947f-82dc-c122-000000000e1d 13731 1727203887.20734: WORKER PROCESS EXITING skipping: [managed-node3] => { "changed": false, "false_condition": "network_state != {}", "skip_reason": "Conditional result was False" } 13731 1727203887.20792: no more pending results, returning what we have 13731 1727203887.20797: results queue empty 13731 1727203887.20797: checking for any_errors_fatal 13731 1727203887.20807: done checking for any_errors_fatal 13731 1727203887.20808: checking for max_fail_percentage 13731 1727203887.20810: done checking for max_fail_percentage 13731 1727203887.20811: checking to see if all hosts have failed and the running result is not ok 13731 1727203887.20812: done checking to see if all hosts have failed 13731 1727203887.20813: getting the remaining hosts for this loop 13731 1727203887.20815: done getting the remaining hosts for this loop 13731 1727203887.20818: getting the next task for host managed-node3 13731 1727203887.20827: done getting next task for host managed-node3 13731 1727203887.20833: ^ task is: TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13731 1727203887.20840: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=22, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203887.20869: getting variables 13731 1727203887.20871: in VariableManager get_vars() 13731 1727203887.21129: Calling all_inventory to load vars for managed-node3 13731 1727203887.21132: Calling groups_inventory to load vars for managed-node3 13731 1727203887.21134: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203887.21143: Calling all_plugins_play to load vars for managed-node3 13731 1727203887.21146: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203887.21149: Calling groups_plugins_play to load vars for managed-node3 13731 1727203887.22594: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203887.24338: done with get_vars() 13731 1727203887.24374: done getting variables 13731 1727203887.24435: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show stderr messages for the network_connections] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:177 Tuesday 24 September 2024 14:51:27 -0400 (0:00:00.056) 0:01:05.475 ***** 13731 1727203887.24481: entering _queue_task() for managed-node3/debug 13731 1727203887.24841: worker is 1 (out of 1 available) 13731 1727203887.24854: exiting _queue_task() for managed-node3/debug 13731 1727203887.24869: done queuing things up, now waiting for results queue to drain 13731 1727203887.24871: waiting for pending results... 13731 1727203887.25286: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections 13731 1727203887.25371: in run() - task 028d2410-947f-82dc-c122-000000000e1e 13731 1727203887.25395: variable 'ansible_search_path' from source: unknown 13731 1727203887.25409: variable 'ansible_search_path' from source: unknown 13731 1727203887.25447: calling self._execute() 13731 1727203887.25558: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.25574: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.25592: variable 'omit' from source: magic vars 13731 1727203887.26174: variable 'ansible_distribution_major_version' from source: facts 13731 1727203887.26181: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203887.26184: variable 'omit' from source: magic vars 13731 1727203887.26229: variable 'omit' from source: magic vars 13731 1727203887.26286: variable 'omit' from source: magic vars 13731 1727203887.26333: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203887.26379: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203887.26412: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203887.26434: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203887.26449: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203887.26581: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203887.26585: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.26587: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.26635: Set connection var ansible_pipelining to False 13731 1727203887.26647: Set connection var ansible_shell_type to sh 13731 1727203887.26657: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203887.26670: Set connection var ansible_connection to ssh 13731 1727203887.26685: Set connection var ansible_shell_executable to /bin/sh 13731 1727203887.26695: Set connection var ansible_timeout to 10 13731 1727203887.26731: variable 'ansible_shell_executable' from source: unknown 13731 1727203887.26739: variable 'ansible_connection' from source: unknown 13731 1727203887.26746: variable 'ansible_module_compression' from source: unknown 13731 1727203887.26753: variable 'ansible_shell_type' from source: unknown 13731 1727203887.26760: variable 'ansible_shell_executable' from source: unknown 13731 1727203887.26770: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.26780: variable 'ansible_pipelining' from source: unknown 13731 1727203887.26787: variable 'ansible_timeout' from source: unknown 13731 1727203887.26795: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.26959: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203887.27042: variable 'omit' from source: magic vars 13731 1727203887.27046: starting attempt loop 13731 1727203887.27048: running the handler 13731 1727203887.27151: variable '__network_connections_result' from source: set_fact 13731 1727203887.27214: handler run complete 13731 1727203887.27235: attempt loop complete, returning result 13731 1727203887.27244: _execute() done 13731 1727203887.27267: dumping result to json 13731 1727203887.27287: done dumping result, returning 13731 1727203887.27482: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show stderr messages for the network_connections [028d2410-947f-82dc-c122-000000000e1e] 13731 1727203887.27486: sending task result for task 028d2410-947f-82dc-c122-000000000e1e 13731 1727203887.27564: done sending task result for task 028d2410-947f-82dc-c122-000000000e1e 13731 1727203887.27568: WORKER PROCESS EXITING ok: [managed-node3] => { "__network_connections_result.stderr_lines": [ "" ] } 13731 1727203887.27653: no more pending results, returning what we have 13731 1727203887.27658: results queue empty 13731 1727203887.27659: checking for any_errors_fatal 13731 1727203887.27670: done checking for any_errors_fatal 13731 1727203887.27671: checking for max_fail_percentage 13731 1727203887.27673: done checking for max_fail_percentage 13731 1727203887.27676: checking to see if all hosts have failed and the running result is not ok 13731 1727203887.27878: done checking to see if all hosts have failed 13731 1727203887.27880: getting the remaining hosts for this loop 13731 1727203887.27882: done getting the remaining hosts for this loop 13731 1727203887.27887: getting the next task for host managed-node3 13731 1727203887.27896: done getting next task for host managed-node3 13731 1727203887.27900: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13731 1727203887.27907: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=23, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203887.27923: getting variables 13731 1727203887.27925: in VariableManager get_vars() 13731 1727203887.28189: Calling all_inventory to load vars for managed-node3 13731 1727203887.28193: Calling groups_inventory to load vars for managed-node3 13731 1727203887.28200: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203887.28211: Calling all_plugins_play to load vars for managed-node3 13731 1727203887.28214: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203887.28217: Calling groups_plugins_play to load vars for managed-node3 13731 1727203887.30408: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203887.32112: done with get_vars() 13731 1727203887.32142: done getting variables 13731 1727203887.32208: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_connections] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:181 Tuesday 24 September 2024 14:51:27 -0400 (0:00:00.077) 0:01:05.553 ***** 13731 1727203887.32257: entering _queue_task() for managed-node3/debug 13731 1727203887.32637: worker is 1 (out of 1 available) 13731 1727203887.32767: exiting _queue_task() for managed-node3/debug 13731 1727203887.32782: done queuing things up, now waiting for results queue to drain 13731 1727203887.32784: waiting for pending results... 13731 1727203887.33095: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections 13731 1727203887.33171: in run() - task 028d2410-947f-82dc-c122-000000000e1f 13731 1727203887.33199: variable 'ansible_search_path' from source: unknown 13731 1727203887.33410: variable 'ansible_search_path' from source: unknown 13731 1727203887.33413: calling self._execute() 13731 1727203887.33472: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.33525: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.33538: variable 'omit' from source: magic vars 13731 1727203887.34340: variable 'ansible_distribution_major_version' from source: facts 13731 1727203887.34358: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203887.34372: variable 'omit' from source: magic vars 13731 1727203887.34468: variable 'omit' from source: magic vars 13731 1727203887.34633: variable 'omit' from source: magic vars 13731 1727203887.34683: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203887.34728: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203887.34752: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203887.34779: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203887.34796: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203887.34833: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203887.34841: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.34848: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.34957: Set connection var ansible_pipelining to False 13731 1727203887.34971: Set connection var ansible_shell_type to sh 13731 1727203887.34982: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203887.34991: Set connection var ansible_connection to ssh 13731 1727203887.34999: Set connection var ansible_shell_executable to /bin/sh 13731 1727203887.35007: Set connection var ansible_timeout to 10 13731 1727203887.35032: variable 'ansible_shell_executable' from source: unknown 13731 1727203887.35150: variable 'ansible_connection' from source: unknown 13731 1727203887.35153: variable 'ansible_module_compression' from source: unknown 13731 1727203887.35155: variable 'ansible_shell_type' from source: unknown 13731 1727203887.35157: variable 'ansible_shell_executable' from source: unknown 13731 1727203887.35159: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.35164: variable 'ansible_pipelining' from source: unknown 13731 1727203887.35166: variable 'ansible_timeout' from source: unknown 13731 1727203887.35168: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.35229: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203887.35245: variable 'omit' from source: magic vars 13731 1727203887.35258: starting attempt loop 13731 1727203887.35268: running the handler 13731 1727203887.35321: variable '__network_connections_result' from source: set_fact 13731 1727203887.35413: variable '__network_connections_result' from source: set_fact 13731 1727203887.35550: handler run complete 13731 1727203887.35590: attempt loop complete, returning result 13731 1727203887.35598: _execute() done 13731 1727203887.35604: dumping result to json 13731 1727203887.35612: done dumping result, returning 13731 1727203887.35624: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_connections [028d2410-947f-82dc-c122-000000000e1f] 13731 1727203887.35633: sending task result for task 028d2410-947f-82dc-c122-000000000e1f ok: [managed-node3] => { "__network_connections_result": { "_invocation": { "module_args": { "__debug_flags": "", "__header": "#\n# Ansible managed\n#\n# system_role:network\n", "connections": [ { "name": "bond0.1", "persistent_state": "absent", "state": "down" }, { "name": "bond0.0", "persistent_state": "absent", "state": "down" }, { "name": "bond0", "persistent_state": "absent", "state": "down" } ], "force_state_change": false, "ignore_errors": false, "provider": "nm" } }, "changed": true, "failed": false, "stderr": "\n", "stderr_lines": [ "" ] } } 13731 1727203887.35977: no more pending results, returning what we have 13731 1727203887.35981: results queue empty 13731 1727203887.35982: checking for any_errors_fatal 13731 1727203887.35991: done checking for any_errors_fatal 13731 1727203887.35991: checking for max_fail_percentage 13731 1727203887.35993: done checking for max_fail_percentage 13731 1727203887.35994: checking to see if all hosts have failed and the running result is not ok 13731 1727203887.35995: done checking to see if all hosts have failed 13731 1727203887.35995: getting the remaining hosts for this loop 13731 1727203887.35997: done getting the remaining hosts for this loop 13731 1727203887.36001: getting the next task for host managed-node3 13731 1727203887.36009: done getting next task for host managed-node3 13731 1727203887.36013: ^ task is: TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13731 1727203887.36018: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=24, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203887.36032: getting variables 13731 1727203887.36034: in VariableManager get_vars() 13731 1727203887.36287: Calling all_inventory to load vars for managed-node3 13731 1727203887.36290: Calling groups_inventory to load vars for managed-node3 13731 1727203887.36292: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203887.36308: Calling all_plugins_play to load vars for managed-node3 13731 1727203887.36310: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203887.36313: Calling groups_plugins_play to load vars for managed-node3 13731 1727203887.36989: done sending task result for task 028d2410-947f-82dc-c122-000000000e1f 13731 1727203887.36993: WORKER PROCESS EXITING 13731 1727203887.37699: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203887.39311: done with get_vars() 13731 1727203887.39337: done getting variables 13731 1727203887.39401: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.network : Show debug messages for the network_state] *** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:186 Tuesday 24 September 2024 14:51:27 -0400 (0:00:00.071) 0:01:05.625 ***** 13731 1727203887.39439: entering _queue_task() for managed-node3/debug 13731 1727203887.39901: worker is 1 (out of 1 available) 13731 1727203887.39911: exiting _queue_task() for managed-node3/debug 13731 1727203887.39923: done queuing things up, now waiting for results queue to drain 13731 1727203887.39924: waiting for pending results... 13731 1727203887.40134: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state 13731 1727203887.40544: in run() - task 028d2410-947f-82dc-c122-000000000e20 13731 1727203887.40569: variable 'ansible_search_path' from source: unknown 13731 1727203887.40578: variable 'ansible_search_path' from source: unknown 13731 1727203887.40622: calling self._execute() 13731 1727203887.40854: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.40869: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.40885: variable 'omit' from source: magic vars 13731 1727203887.41656: variable 'ansible_distribution_major_version' from source: facts 13731 1727203887.41713: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203887.42133: variable 'network_state' from source: role '' defaults 13731 1727203887.42136: Evaluated conditional (network_state != {}): False 13731 1727203887.42138: when evaluation is False, skipping this task 13731 1727203887.42141: _execute() done 13731 1727203887.42143: dumping result to json 13731 1727203887.42145: done dumping result, returning 13731 1727203887.42148: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Show debug messages for the network_state [028d2410-947f-82dc-c122-000000000e20] 13731 1727203887.42150: sending task result for task 028d2410-947f-82dc-c122-000000000e20 13731 1727203887.42224: done sending task result for task 028d2410-947f-82dc-c122-000000000e20 13731 1727203887.42227: WORKER PROCESS EXITING skipping: [managed-node3] => { "false_condition": "network_state != {}" } 13731 1727203887.42288: no more pending results, returning what we have 13731 1727203887.42294: results queue empty 13731 1727203887.42294: checking for any_errors_fatal 13731 1727203887.42306: done checking for any_errors_fatal 13731 1727203887.42307: checking for max_fail_percentage 13731 1727203887.42309: done checking for max_fail_percentage 13731 1727203887.42310: checking to see if all hosts have failed and the running result is not ok 13731 1727203887.42310: done checking to see if all hosts have failed 13731 1727203887.42311: getting the remaining hosts for this loop 13731 1727203887.42314: done getting the remaining hosts for this loop 13731 1727203887.42317: getting the next task for host managed-node3 13731 1727203887.42328: done getting next task for host managed-node3 13731 1727203887.42332: ^ task is: TASK: fedora.linux_system_roles.network : Re-test connectivity 13731 1727203887.42338: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=25, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203887.42366: getting variables 13731 1727203887.42368: in VariableManager get_vars() 13731 1727203887.42416: Calling all_inventory to load vars for managed-node3 13731 1727203887.42419: Calling groups_inventory to load vars for managed-node3 13731 1727203887.42421: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203887.42434: Calling all_plugins_play to load vars for managed-node3 13731 1727203887.42437: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203887.42440: Calling groups_plugins_play to load vars for managed-node3 13731 1727203887.45692: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203887.47294: done with get_vars() 13731 1727203887.47319: done getting variables TASK [fedora.linux_system_roles.network : Re-test connectivity] **************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:192 Tuesday 24 September 2024 14:51:27 -0400 (0:00:00.079) 0:01:05.705 ***** 13731 1727203887.47422: entering _queue_task() for managed-node3/ping 13731 1727203887.47784: worker is 1 (out of 1 available) 13731 1727203887.47795: exiting _queue_task() for managed-node3/ping 13731 1727203887.47807: done queuing things up, now waiting for results queue to drain 13731 1727203887.47808: waiting for pending results... 13731 1727203887.48110: running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Re-test connectivity 13731 1727203887.48272: in run() - task 028d2410-947f-82dc-c122-000000000e21 13731 1727203887.48297: variable 'ansible_search_path' from source: unknown 13731 1727203887.48381: variable 'ansible_search_path' from source: unknown 13731 1727203887.48384: calling self._execute() 13731 1727203887.48453: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.48530: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.48545: variable 'omit' from source: magic vars 13731 1727203887.49353: variable 'ansible_distribution_major_version' from source: facts 13731 1727203887.49402: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203887.49414: variable 'omit' from source: magic vars 13731 1727203887.49570: variable 'omit' from source: magic vars 13731 1727203887.49833: variable 'omit' from source: magic vars 13731 1727203887.49836: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203887.49887: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203887.49914: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203887.49941: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203887.49995: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203887.50031: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203887.50268: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.50271: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.50273: Set connection var ansible_pipelining to False 13731 1727203887.50275: Set connection var ansible_shell_type to sh 13731 1727203887.50482: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203887.50485: Set connection var ansible_connection to ssh 13731 1727203887.50488: Set connection var ansible_shell_executable to /bin/sh 13731 1727203887.50490: Set connection var ansible_timeout to 10 13731 1727203887.50493: variable 'ansible_shell_executable' from source: unknown 13731 1727203887.50495: variable 'ansible_connection' from source: unknown 13731 1727203887.50497: variable 'ansible_module_compression' from source: unknown 13731 1727203887.50499: variable 'ansible_shell_type' from source: unknown 13731 1727203887.50501: variable 'ansible_shell_executable' from source: unknown 13731 1727203887.50503: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.50505: variable 'ansible_pipelining' from source: unknown 13731 1727203887.50507: variable 'ansible_timeout' from source: unknown 13731 1727203887.50509: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.51143: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 13731 1727203887.51163: variable 'omit' from source: magic vars 13731 1727203887.51174: starting attempt loop 13731 1727203887.51184: running the handler 13731 1727203887.51206: _low_level_execute_command(): starting 13731 1727203887.51218: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203887.52196: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.52265: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203887.52288: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.52312: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.52466: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.54135: stdout chunk (state=3): >>>/root <<< 13731 1727203887.54283: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.54287: stdout chunk (state=3): >>><<< 13731 1727203887.54289: stderr chunk (state=3): >>><<< 13731 1727203887.54503: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203887.54507: _low_level_execute_command(): starting 13731 1727203887.54510: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618 `" && echo ansible-tmp-1727203887.5440674-18339-82123234324618="` echo /root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618 `" ) && sleep 0' 13731 1727203887.55503: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.55516: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.55529: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.55684: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.55991: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.56055: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.57949: stdout chunk (state=3): >>>ansible-tmp-1727203887.5440674-18339-82123234324618=/root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618 <<< 13731 1727203887.58067: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.58070: stderr chunk (state=3): >>><<< 13731 1727203887.58073: stdout chunk (state=3): >>><<< 13731 1727203887.58383: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203887.5440674-18339-82123234324618=/root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203887.58387: variable 'ansible_module_compression' from source: unknown 13731 1727203887.58389: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.ping-ZIP_DEFLATED 13731 1727203887.58391: variable 'ansible_facts' from source: unknown 13731 1727203887.58542: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/AnsiballZ_ping.py 13731 1727203887.59157: Sending initial data 13731 1727203887.59163: Sent initial data (152 bytes) 13731 1727203887.60166: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.60172: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203887.60216: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.60222: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203887.60257: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.60265: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203887.60320: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.60507: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203887.60511: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.60534: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.60710: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.62099: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203887.62197: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203887.62201: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpd8ggr17c /root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/AnsiballZ_ping.py <<< 13731 1727203887.62204: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/AnsiballZ_ping.py" <<< 13731 1727203887.62258: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpd8ggr17c" to remote "/root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/AnsiballZ_ping.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/AnsiballZ_ping.py" <<< 13731 1727203887.62834: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.62882: stderr chunk (state=3): >>><<< 13731 1727203887.62885: stdout chunk (state=3): >>><<< 13731 1727203887.62903: done transferring module to remote 13731 1727203887.62913: _low_level_execute_command(): starting 13731 1727203887.62917: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/ /root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/AnsiballZ_ping.py && sleep 0' 13731 1727203887.63342: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.63346: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.63348: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.63350: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203887.63352: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.63407: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.63410: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.63437: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.65226: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.65230: stdout chunk (state=3): >>><<< 13731 1727203887.65232: stderr chunk (state=3): >>><<< 13731 1727203887.65282: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203887.65285: _low_level_execute_command(): starting 13731 1727203887.65287: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/AnsiballZ_ping.py && sleep 0' 13731 1727203887.65732: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203887.65735: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.65738: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203887.65740: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203887.65742: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.65782: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.65802: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.65838: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.80516: stdout chunk (state=3): >>> {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} <<< 13731 1727203887.81710: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203887.81737: stderr chunk (state=3): >>><<< 13731 1727203887.81740: stdout chunk (state=3): >>><<< 13731 1727203887.81756: _low_level_execute_command() done: rc=0, stdout= {"ping": "pong", "invocation": {"module_args": {"data": "pong"}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203887.81866: done with _execute_module (ping, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ping', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203887.81874: _low_level_execute_command(): starting 13731 1727203887.81879: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203887.5440674-18339-82123234324618/ > /dev/null 2>&1 && sleep 0' 13731 1727203887.82614: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203887.82626: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.82637: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203887.82653: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203887.82670: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203887.82680: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203887.82693: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.82714: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203887.82781: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203887.82797: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.82831: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203887.82850: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.82881: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.82954: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.84789: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.84830: stdout chunk (state=3): >>><<< 13731 1727203887.84834: stderr chunk (state=3): >>><<< 13731 1727203887.84852: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203887.84983: handler run complete 13731 1727203887.84987: attempt loop complete, returning result 13731 1727203887.84990: _execute() done 13731 1727203887.84992: dumping result to json 13731 1727203887.84994: done dumping result, returning 13731 1727203887.84996: done running TaskExecutor() for managed-node3/TASK: fedora.linux_system_roles.network : Re-test connectivity [028d2410-947f-82dc-c122-000000000e21] 13731 1727203887.84998: sending task result for task 028d2410-947f-82dc-c122-000000000e21 13731 1727203887.85069: done sending task result for task 028d2410-947f-82dc-c122-000000000e21 13731 1727203887.85072: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "ping": "pong" } 13731 1727203887.85147: no more pending results, returning what we have 13731 1727203887.85152: results queue empty 13731 1727203887.85153: checking for any_errors_fatal 13731 1727203887.85163: done checking for any_errors_fatal 13731 1727203887.85164: checking for max_fail_percentage 13731 1727203887.85166: done checking for max_fail_percentage 13731 1727203887.85167: checking to see if all hosts have failed and the running result is not ok 13731 1727203887.85168: done checking to see if all hosts have failed 13731 1727203887.85168: getting the remaining hosts for this loop 13731 1727203887.85170: done getting the remaining hosts for this loop 13731 1727203887.85174: getting the next task for host managed-node3 13731 1727203887.85190: done getting next task for host managed-node3 13731 1727203887.85193: ^ task is: TASK: meta (role_complete) 13731 1727203887.85199: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203887.85214: getting variables 13731 1727203887.85216: in VariableManager get_vars() 13731 1727203887.85268: Calling all_inventory to load vars for managed-node3 13731 1727203887.85272: Calling groups_inventory to load vars for managed-node3 13731 1727203887.85274: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203887.85491: Calling all_plugins_play to load vars for managed-node3 13731 1727203887.85502: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203887.85506: Calling groups_plugins_play to load vars for managed-node3 13731 1727203887.86482: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203887.87371: done with get_vars() 13731 1727203887.87393: done getting variables 13731 1727203887.87457: done queuing things up, now waiting for results queue to drain 13731 1727203887.87459: results queue empty 13731 1727203887.87459: checking for any_errors_fatal 13731 1727203887.87463: done checking for any_errors_fatal 13731 1727203887.87464: checking for max_fail_percentage 13731 1727203887.87465: done checking for max_fail_percentage 13731 1727203887.87465: checking to see if all hosts have failed and the running result is not ok 13731 1727203887.87466: done checking to see if all hosts have failed 13731 1727203887.87466: getting the remaining hosts for this loop 13731 1727203887.87467: done getting the remaining hosts for this loop 13731 1727203887.87469: getting the next task for host managed-node3 13731 1727203887.87472: done getting next task for host managed-node3 13731 1727203887.87474: ^ task is: TASK: Delete the device '{{ controller_device }}' 13731 1727203887.87477: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203887.87479: getting variables 13731 1727203887.87480: in VariableManager get_vars() 13731 1727203887.87493: Calling all_inventory to load vars for managed-node3 13731 1727203887.87494: Calling groups_inventory to load vars for managed-node3 13731 1727203887.87495: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203887.87499: Calling all_plugins_play to load vars for managed-node3 13731 1727203887.87500: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203887.87502: Calling groups_plugins_play to load vars for managed-node3 13731 1727203887.88234: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203887.89101: done with get_vars() 13731 1727203887.89118: done getting variables 13731 1727203887.89150: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) 13731 1727203887.89242: variable 'controller_device' from source: play vars TASK [Delete the device 'nm-bond'] ********************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_bond_profile+device.yml:22 Tuesday 24 September 2024 14:51:27 -0400 (0:00:00.418) 0:01:06.123 ***** 13731 1727203887.89273: entering _queue_task() for managed-node3/command 13731 1727203887.89595: worker is 1 (out of 1 available) 13731 1727203887.89609: exiting _queue_task() for managed-node3/command 13731 1727203887.89621: done queuing things up, now waiting for results queue to drain 13731 1727203887.89623: waiting for pending results... 13731 1727203887.89810: running TaskExecutor() for managed-node3/TASK: Delete the device 'nm-bond' 13731 1727203887.89900: in run() - task 028d2410-947f-82dc-c122-000000000e51 13731 1727203887.89912: variable 'ansible_search_path' from source: unknown 13731 1727203887.89916: variable 'ansible_search_path' from source: unknown 13731 1727203887.89943: calling self._execute() 13731 1727203887.90024: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.90029: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.90037: variable 'omit' from source: magic vars 13731 1727203887.90337: variable 'ansible_distribution_major_version' from source: facts 13731 1727203887.90347: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203887.90352: variable 'omit' from source: magic vars 13731 1727203887.90371: variable 'omit' from source: magic vars 13731 1727203887.90438: variable 'controller_device' from source: play vars 13731 1727203887.90452: variable 'omit' from source: magic vars 13731 1727203887.90487: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203887.90520: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203887.90532: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203887.90545: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203887.90555: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203887.90581: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203887.90584: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.90586: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.90659: Set connection var ansible_pipelining to False 13731 1727203887.90665: Set connection var ansible_shell_type to sh 13731 1727203887.90667: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203887.90673: Set connection var ansible_connection to ssh 13731 1727203887.90679: Set connection var ansible_shell_executable to /bin/sh 13731 1727203887.90684: Set connection var ansible_timeout to 10 13731 1727203887.90701: variable 'ansible_shell_executable' from source: unknown 13731 1727203887.90704: variable 'ansible_connection' from source: unknown 13731 1727203887.90706: variable 'ansible_module_compression' from source: unknown 13731 1727203887.90710: variable 'ansible_shell_type' from source: unknown 13731 1727203887.90712: variable 'ansible_shell_executable' from source: unknown 13731 1727203887.90714: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203887.90717: variable 'ansible_pipelining' from source: unknown 13731 1727203887.90719: variable 'ansible_timeout' from source: unknown 13731 1727203887.90724: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203887.90825: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203887.90837: variable 'omit' from source: magic vars 13731 1727203887.90840: starting attempt loop 13731 1727203887.90843: running the handler 13731 1727203887.90858: _low_level_execute_command(): starting 13731 1727203887.90865: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203887.91396: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203887.91400: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.91403: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203887.91407: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.91456: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203887.91459: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.91462: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.91506: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.93062: stdout chunk (state=3): >>>/root <<< 13731 1727203887.93156: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.93188: stderr chunk (state=3): >>><<< 13731 1727203887.93191: stdout chunk (state=3): >>><<< 13731 1727203887.93212: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203887.93224: _low_level_execute_command(): starting 13731 1727203887.93231: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093 `" && echo ansible-tmp-1727203887.9321249-18370-270577168834093="` echo /root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093 `" ) && sleep 0' 13731 1727203887.93677: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.93681: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203887.93691: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203887.93695: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.93742: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203887.93749: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.93751: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.93784: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.95618: stdout chunk (state=3): >>>ansible-tmp-1727203887.9321249-18370-270577168834093=/root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093 <<< 13731 1727203887.95886: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.95889: stdout chunk (state=3): >>><<< 13731 1727203887.95892: stderr chunk (state=3): >>><<< 13731 1727203887.95894: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203887.9321249-18370-270577168834093=/root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203887.95897: variable 'ansible_module_compression' from source: unknown 13731 1727203887.95899: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203887.95938: variable 'ansible_facts' from source: unknown 13731 1727203887.96041: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/AnsiballZ_command.py 13731 1727203887.96226: Sending initial data 13731 1727203887.96235: Sent initial data (156 bytes) 13731 1727203887.96852: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203887.96878: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203887.96893: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203887.96987: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203887.97017: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203887.97036: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203887.97062: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203887.97134: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203887.98635: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203887.98694: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203887.98740: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpkerz_wqn /root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/AnsiballZ_command.py <<< 13731 1727203887.98772: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/AnsiballZ_command.py" <<< 13731 1727203887.98792: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpkerz_wqn" to remote "/root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/AnsiballZ_command.py" <<< 13731 1727203887.99481: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203887.99533: stderr chunk (state=3): >>><<< 13731 1727203887.99541: stdout chunk (state=3): >>><<< 13731 1727203887.99606: done transferring module to remote 13731 1727203887.99621: _low_level_execute_command(): starting 13731 1727203887.99636: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/ /root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/AnsiballZ_command.py && sleep 0' 13731 1727203888.00404: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.00408: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.00426: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.00440: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.00509: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.02206: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.02266: stderr chunk (state=3): >>><<< 13731 1727203888.02281: stdout chunk (state=3): >>><<< 13731 1727203888.02302: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.02309: _low_level_execute_command(): starting 13731 1727203888.02318: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/AnsiballZ_command.py && sleep 0' 13731 1727203888.02956: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203888.02973: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.02990: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.03006: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203888.03022: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203888.03096: stderr chunk (state=3): >>>debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.03151: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.03166: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.03189: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.03271: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.18949: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "Cannot find device \"nm-bond\"", "rc": 1, "cmd": ["ip", "link", "del", "nm-bond"], "start": "2024-09-24 14:51:28.181699", "end": "2024-09-24 14:51:28.188656", "delta": "0:00:00.006957", "failed": true, "msg": "non-zero return code", "invocation": {"module_args": {"_raw_params": "ip link del nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203888.20244: stderr chunk (state=3): >>>debug2: Received exit status from master 1 Shared connection to 10.31.47.22 closed. <<< 13731 1727203888.20278: stderr chunk (state=3): >>><<< 13731 1727203888.20281: stdout chunk (state=3): >>><<< 13731 1727203888.20299: _low_level_execute_command() done: rc=1, stdout= {"changed": true, "stdout": "", "stderr": "Cannot find device \"nm-bond\"", "rc": 1, "cmd": ["ip", "link", "del", "nm-bond"], "start": "2024-09-24 14:51:28.181699", "end": "2024-09-24 14:51:28.188656", "delta": "0:00:00.006957", "failed": true, "msg": "non-zero return code", "invocation": {"module_args": {"_raw_params": "ip link del nm-bond", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 1 Shared connection to 10.31.47.22 closed. 13731 1727203888.20329: done with _execute_module (ansible.legacy.command, {'_raw_params': 'ip link del nm-bond', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203888.20337: _low_level_execute_command(): starting 13731 1727203888.20342: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203887.9321249-18370-270577168834093/ > /dev/null 2>&1 && sleep 0' 13731 1727203888.20767: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.20771: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.20792: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203888.20795: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.20811: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.20877: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.20881: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.20885: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.20920: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.22733: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.22757: stderr chunk (state=3): >>><<< 13731 1727203888.22762: stdout chunk (state=3): >>><<< 13731 1727203888.22774: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.22784: handler run complete 13731 1727203888.22802: Evaluated conditional (False): False 13731 1727203888.22805: Evaluated conditional (False): False 13731 1727203888.22814: attempt loop complete, returning result 13731 1727203888.22816: _execute() done 13731 1727203888.22819: dumping result to json 13731 1727203888.22823: done dumping result, returning 13731 1727203888.22831: done running TaskExecutor() for managed-node3/TASK: Delete the device 'nm-bond' [028d2410-947f-82dc-c122-000000000e51] 13731 1727203888.22837: sending task result for task 028d2410-947f-82dc-c122-000000000e51 13731 1727203888.22939: done sending task result for task 028d2410-947f-82dc-c122-000000000e51 13731 1727203888.22942: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": [ "ip", "link", "del", "nm-bond" ], "delta": "0:00:00.006957", "end": "2024-09-24 14:51:28.188656", "failed_when_result": false, "rc": 1, "start": "2024-09-24 14:51:28.181699" } STDERR: Cannot find device "nm-bond" MSG: non-zero return code 13731 1727203888.23009: no more pending results, returning what we have 13731 1727203888.23013: results queue empty 13731 1727203888.23014: checking for any_errors_fatal 13731 1727203888.23015: done checking for any_errors_fatal 13731 1727203888.23016: checking for max_fail_percentage 13731 1727203888.23018: done checking for max_fail_percentage 13731 1727203888.23018: checking to see if all hosts have failed and the running result is not ok 13731 1727203888.23019: done checking to see if all hosts have failed 13731 1727203888.23020: getting the remaining hosts for this loop 13731 1727203888.23022: done getting the remaining hosts for this loop 13731 1727203888.23025: getting the next task for host managed-node3 13731 1727203888.23036: done getting next task for host managed-node3 13731 1727203888.23038: ^ task is: TASK: Remove test interfaces 13731 1727203888.23042: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203888.23047: getting variables 13731 1727203888.23048: in VariableManager get_vars() 13731 1727203888.23099: Calling all_inventory to load vars for managed-node3 13731 1727203888.23102: Calling groups_inventory to load vars for managed-node3 13731 1727203888.23105: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203888.23115: Calling all_plugins_play to load vars for managed-node3 13731 1727203888.23118: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203888.23120: Calling groups_plugins_play to load vars for managed-node3 13731 1727203888.23962: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203888.24943: done with get_vars() 13731 1727203888.24959: done getting variables 13731 1727203888.25007: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Remove test interfaces] ************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml:3 Tuesday 24 September 2024 14:51:28 -0400 (0:00:00.357) 0:01:06.481 ***** 13731 1727203888.25032: entering _queue_task() for managed-node3/shell 13731 1727203888.25297: worker is 1 (out of 1 available) 13731 1727203888.25311: exiting _queue_task() for managed-node3/shell 13731 1727203888.25324: done queuing things up, now waiting for results queue to drain 13731 1727203888.25326: waiting for pending results... 13731 1727203888.25509: running TaskExecutor() for managed-node3/TASK: Remove test interfaces 13731 1727203888.25596: in run() - task 028d2410-947f-82dc-c122-000000000e57 13731 1727203888.25609: variable 'ansible_search_path' from source: unknown 13731 1727203888.25612: variable 'ansible_search_path' from source: unknown 13731 1727203888.25640: calling self._execute() 13731 1727203888.25718: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203888.25722: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203888.25731: variable 'omit' from source: magic vars 13731 1727203888.26010: variable 'ansible_distribution_major_version' from source: facts 13731 1727203888.26020: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203888.26026: variable 'omit' from source: magic vars 13731 1727203888.26057: variable 'omit' from source: magic vars 13731 1727203888.26170: variable 'dhcp_interface1' from source: play vars 13731 1727203888.26174: variable 'dhcp_interface2' from source: play vars 13731 1727203888.26191: variable 'omit' from source: magic vars 13731 1727203888.26228: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203888.26254: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203888.26271: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203888.26287: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203888.26297: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203888.26324: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203888.26328: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203888.26330: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203888.26401: Set connection var ansible_pipelining to False 13731 1727203888.26405: Set connection var ansible_shell_type to sh 13731 1727203888.26412: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203888.26415: Set connection var ansible_connection to ssh 13731 1727203888.26420: Set connection var ansible_shell_executable to /bin/sh 13731 1727203888.26427: Set connection var ansible_timeout to 10 13731 1727203888.26481: variable 'ansible_shell_executable' from source: unknown 13731 1727203888.26484: variable 'ansible_connection' from source: unknown 13731 1727203888.26486: variable 'ansible_module_compression' from source: unknown 13731 1727203888.26488: variable 'ansible_shell_type' from source: unknown 13731 1727203888.26490: variable 'ansible_shell_executable' from source: unknown 13731 1727203888.26492: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203888.26494: variable 'ansible_pipelining' from source: unknown 13731 1727203888.26496: variable 'ansible_timeout' from source: unknown 13731 1727203888.26499: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203888.26567: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203888.26576: variable 'omit' from source: magic vars 13731 1727203888.26582: starting attempt loop 13731 1727203888.26585: running the handler 13731 1727203888.26595: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203888.26610: _low_level_execute_command(): starting 13731 1727203888.26617: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203888.27155: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.27163: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203888.27166: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203888.27169: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.27221: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.27224: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.27228: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.27262: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.28812: stdout chunk (state=3): >>>/root <<< 13731 1727203888.28910: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.28942: stderr chunk (state=3): >>><<< 13731 1727203888.28947: stdout chunk (state=3): >>><<< 13731 1727203888.28968: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.28981: _low_level_execute_command(): starting 13731 1727203888.28987: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902 `" && echo ansible-tmp-1727203888.2896805-18387-55538120871902="` echo /root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902 `" ) && sleep 0' 13731 1727203888.29439: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.29451: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203888.29455: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.29457: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.29460: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.29504: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.29507: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.29548: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.31360: stdout chunk (state=3): >>>ansible-tmp-1727203888.2896805-18387-55538120871902=/root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902 <<< 13731 1727203888.31468: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.31495: stderr chunk (state=3): >>><<< 13731 1727203888.31498: stdout chunk (state=3): >>><<< 13731 1727203888.31514: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203888.2896805-18387-55538120871902=/root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.31544: variable 'ansible_module_compression' from source: unknown 13731 1727203888.31588: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203888.31617: variable 'ansible_facts' from source: unknown 13731 1727203888.31678: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/AnsiballZ_command.py 13731 1727203888.31781: Sending initial data 13731 1727203888.31784: Sent initial data (155 bytes) 13731 1727203888.32231: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.32234: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203888.32236: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.32238: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.32240: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.32292: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.32295: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.32330: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.33824: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 <<< 13731 1727203888.33827: stderr chunk (state=3): >>>debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203888.33856: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203888.33890: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpxmu3cfog /root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/AnsiballZ_command.py <<< 13731 1727203888.33893: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/AnsiballZ_command.py" <<< 13731 1727203888.33921: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpxmu3cfog" to remote "/root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/AnsiballZ_command.py" <<< 13731 1727203888.34435: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.34473: stderr chunk (state=3): >>><<< 13731 1727203888.34478: stdout chunk (state=3): >>><<< 13731 1727203888.34494: done transferring module to remote 13731 1727203888.34504: _low_level_execute_command(): starting 13731 1727203888.34512: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/ /root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/AnsiballZ_command.py && sleep 0' 13731 1727203888.34943: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.34947: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203888.34953: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.34955: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.34957: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.35004: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.35007: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.35044: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.36737: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.36760: stderr chunk (state=3): >>><<< 13731 1727203888.36763: stdout chunk (state=3): >>><<< 13731 1727203888.36781: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.36786: _low_level_execute_command(): starting 13731 1727203888.36793: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/AnsiballZ_command.py && sleep 0' 13731 1727203888.37225: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.37229: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203888.37232: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.37234: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.37236: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.37286: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.37289: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.37328: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.56809: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ rc=0\n+ ip link delete test1\n+ '[' 0 '!=' 0 ']'\n+ ip link delete test2\n+ '[' 0 '!=' 0 ']'\n+ ip link delete testbr\n+ '[' 0 '!=' 0 ']'", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "start": "2024-09-24 14:51:28.520845", "end": "2024-09-24 14:51:28.567603", "delta": "0:00:00.046758", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}}<<< 13731 1727203888.56977: stdout chunk (state=3): >>> <<< 13731 1727203888.58417: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203888.58448: stderr chunk (state=3): >>><<< 13731 1727203888.58452: stdout chunk (state=3): >>><<< 13731 1727203888.58469: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ rc=0\n+ ip link delete test1\n+ '[' 0 '!=' 0 ']'\n+ ip link delete test2\n+ '[' 0 '!=' 0 ']'\n+ ip link delete testbr\n+ '[' 0 '!=' 0 ']'", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "start": "2024-09-24 14:51:28.520845", "end": "2024-09-24 14:51:28.567603", "delta": "0:00:00.046758", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203888.58505: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link test1 - error "$rc"\nfi\nip link delete test2 || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link test2 - error "$rc"\nfi\nip link delete testbr || rc="$?"\nif [ "$rc" != 0 ]; then\n echo ERROR - could not delete link testbr - error "$rc"\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203888.58512: _low_level_execute_command(): starting 13731 1727203888.58517: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203888.2896805-18387-55538120871902/ > /dev/null 2>&1 && sleep 0' 13731 1727203888.58957: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.58995: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203888.58998: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203888.59000: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.59002: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.59004: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.59059: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.59062: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.59079: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.59100: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.60906: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.60930: stderr chunk (state=3): >>><<< 13731 1727203888.60933: stdout chunk (state=3): >>><<< 13731 1727203888.60946: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.60953: handler run complete 13731 1727203888.60973: Evaluated conditional (False): False 13731 1727203888.60984: attempt loop complete, returning result 13731 1727203888.60987: _execute() done 13731 1727203888.60991: dumping result to json 13731 1727203888.60999: done dumping result, returning 13731 1727203888.61007: done running TaskExecutor() for managed-node3/TASK: Remove test interfaces [028d2410-947f-82dc-c122-000000000e57] 13731 1727203888.61012: sending task result for task 028d2410-947f-82dc-c122-000000000e57 13731 1727203888.61106: done sending task result for task 028d2410-947f-82dc-c122-000000000e57 13731 1727203888.61110: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nip link delete test1 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test1 - error \"$rc\"\nfi\nip link delete test2 || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link test2 - error \"$rc\"\nfi\nip link delete testbr || rc=\"$?\"\nif [ \"$rc\" != 0 ]; then\n echo ERROR - could not delete link testbr - error \"$rc\"\nfi\n", "delta": "0:00:00.046758", "end": "2024-09-24 14:51:28.567603", "rc": 0, "start": "2024-09-24 14:51:28.520845" } STDERR: + exec + rc=0 + ip link delete test1 + '[' 0 '!=' 0 ']' + ip link delete test2 + '[' 0 '!=' 0 ']' + ip link delete testbr + '[' 0 '!=' 0 ']' 13731 1727203888.61172: no more pending results, returning what we have 13731 1727203888.61177: results queue empty 13731 1727203888.61178: checking for any_errors_fatal 13731 1727203888.61191: done checking for any_errors_fatal 13731 1727203888.61192: checking for max_fail_percentage 13731 1727203888.61193: done checking for max_fail_percentage 13731 1727203888.61194: checking to see if all hosts have failed and the running result is not ok 13731 1727203888.61195: done checking to see if all hosts have failed 13731 1727203888.61195: getting the remaining hosts for this loop 13731 1727203888.61197: done getting the remaining hosts for this loop 13731 1727203888.61200: getting the next task for host managed-node3 13731 1727203888.61209: done getting next task for host managed-node3 13731 1727203888.61211: ^ task is: TASK: Stop dnsmasq/radvd services 13731 1727203888.61216: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203888.61226: getting variables 13731 1727203888.61228: in VariableManager get_vars() 13731 1727203888.61274: Calling all_inventory to load vars for managed-node3 13731 1727203888.61280: Calling groups_inventory to load vars for managed-node3 13731 1727203888.61283: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203888.61297: Calling all_plugins_play to load vars for managed-node3 13731 1727203888.61300: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203888.61302: Calling groups_plugins_play to load vars for managed-node3 13731 1727203888.62111: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203888.62966: done with get_vars() 13731 1727203888.62984: done getting variables 13731 1727203888.63027: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Stop dnsmasq/radvd services] ********************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/remove_test_interfaces_with_dhcp.yml:23 Tuesday 24 September 2024 14:51:28 -0400 (0:00:00.380) 0:01:06.861 ***** 13731 1727203888.63049: entering _queue_task() for managed-node3/shell 13731 1727203888.63297: worker is 1 (out of 1 available) 13731 1727203888.63311: exiting _queue_task() for managed-node3/shell 13731 1727203888.63323: done queuing things up, now waiting for results queue to drain 13731 1727203888.63325: waiting for pending results... 13731 1727203888.63515: running TaskExecutor() for managed-node3/TASK: Stop dnsmasq/radvd services 13731 1727203888.63592: in run() - task 028d2410-947f-82dc-c122-000000000e58 13731 1727203888.63604: variable 'ansible_search_path' from source: unknown 13731 1727203888.63607: variable 'ansible_search_path' from source: unknown 13731 1727203888.63636: calling self._execute() 13731 1727203888.63715: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203888.63719: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203888.63727: variable 'omit' from source: magic vars 13731 1727203888.64014: variable 'ansible_distribution_major_version' from source: facts 13731 1727203888.64024: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203888.64029: variable 'omit' from source: magic vars 13731 1727203888.64068: variable 'omit' from source: magic vars 13731 1727203888.64095: variable 'omit' from source: magic vars 13731 1727203888.64128: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203888.64154: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203888.64173: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203888.64188: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203888.64200: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203888.64228: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203888.64231: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203888.64233: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203888.64305: Set connection var ansible_pipelining to False 13731 1727203888.64308: Set connection var ansible_shell_type to sh 13731 1727203888.64315: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203888.64318: Set connection var ansible_connection to ssh 13731 1727203888.64330: Set connection var ansible_shell_executable to /bin/sh 13731 1727203888.64333: Set connection var ansible_timeout to 10 13731 1727203888.64348: variable 'ansible_shell_executable' from source: unknown 13731 1727203888.64351: variable 'ansible_connection' from source: unknown 13731 1727203888.64353: variable 'ansible_module_compression' from source: unknown 13731 1727203888.64356: variable 'ansible_shell_type' from source: unknown 13731 1727203888.64358: variable 'ansible_shell_executable' from source: unknown 13731 1727203888.64360: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203888.64366: variable 'ansible_pipelining' from source: unknown 13731 1727203888.64369: variable 'ansible_timeout' from source: unknown 13731 1727203888.64372: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203888.64480: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203888.64489: variable 'omit' from source: magic vars 13731 1727203888.64494: starting attempt loop 13731 1727203888.64497: running the handler 13731 1727203888.64506: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203888.64521: _low_level_execute_command(): starting 13731 1727203888.64528: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203888.65058: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.65062: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.65065: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203888.65067: stderr chunk (state=3): >>>debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.65123: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.65127: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.65131: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.65164: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.66728: stdout chunk (state=3): >>>/root <<< 13731 1727203888.66825: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.66859: stderr chunk (state=3): >>><<< 13731 1727203888.66862: stdout chunk (state=3): >>><<< 13731 1727203888.66886: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.66897: _low_level_execute_command(): starting 13731 1727203888.66903: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994 `" && echo ansible-tmp-1727203888.668858-18395-167751050130994="` echo /root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994 `" ) && sleep 0' 13731 1727203888.67346: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.67349: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203888.67362: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass <<< 13731 1727203888.67365: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.67368: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.67410: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.67420: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.67449: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.69288: stdout chunk (state=3): >>>ansible-tmp-1727203888.668858-18395-167751050130994=/root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994 <<< 13731 1727203888.69392: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.69416: stderr chunk (state=3): >>><<< 13731 1727203888.69419: stdout chunk (state=3): >>><<< 13731 1727203888.69435: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203888.668858-18395-167751050130994=/root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.69465: variable 'ansible_module_compression' from source: unknown 13731 1727203888.69510: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203888.69537: variable 'ansible_facts' from source: unknown 13731 1727203888.69598: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/AnsiballZ_command.py 13731 1727203888.69704: Sending initial data 13731 1727203888.69707: Sent initial data (155 bytes) 13731 1727203888.70133: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.70169: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203888.70172: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203888.70174: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.70179: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.70181: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found <<< 13731 1727203888.70183: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.70230: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.70237: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.70274: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.71779: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203888.71809: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203888.71846: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp82pefi6h /root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/AnsiballZ_command.py <<< 13731 1727203888.71848: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/AnsiballZ_command.py" <<< 13731 1727203888.71877: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory <<< 13731 1727203888.71882: stderr chunk (state=3): >>>debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp82pefi6h" to remote "/root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/AnsiballZ_command.py" <<< 13731 1727203888.72389: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.72430: stderr chunk (state=3): >>><<< 13731 1727203888.72434: stdout chunk (state=3): >>><<< 13731 1727203888.72478: done transferring module to remote 13731 1727203888.72488: _low_level_execute_command(): starting 13731 1727203888.72495: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/ /root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/AnsiballZ_command.py && sleep 0' 13731 1727203888.72949: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.72952: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.72954: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203888.72960: stderr chunk (state=3): >>>debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.73003: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.73016: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.73059: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.74749: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.74777: stderr chunk (state=3): >>><<< 13731 1727203888.74780: stdout chunk (state=3): >>><<< 13731 1727203888.74794: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.74798: _low_level_execute_command(): starting 13731 1727203888.74802: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/AnsiballZ_command.py && sleep 0' 13731 1727203888.75258: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203888.75261: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203888.75264: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.75266: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203888.75268: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.75326: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.75344: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.75348: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.75370: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.92938: stdout chunk (state=3): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ pkill -F /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.lease\n+ grep 'release 6' /etc/redhat-release\n+ systemctl is-active firewalld\ninactive", "rc": 0, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "start": "2024-09-24 14:51:28.902560", "end": "2024-09-24 14:51:28.928183", "delta": "0:00:00.025623", "msg": "", "invocation": {"module_args": {"_raw_params": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203888.94398: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203888.94402: stdout chunk (state=3): >>><<< 13731 1727203888.94405: stderr chunk (state=3): >>><<< 13731 1727203888.94424: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ pkill -F /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.pid\n+ rm -rf /run/dhcp_testbr.lease\n+ grep 'release 6' /etc/redhat-release\n+ systemctl is-active firewalld\ninactive", "rc": 0, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "start": "2024-09-24 14:51:28.902560", "end": "2024-09-24 14:51:28.928183", "delta": "0:00:00.025623", "msg": "", "invocation": {"module_args": {"_raw_params": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203888.94478: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep \'release 6\' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service="$service"; then\n firewall-cmd --remove-service "$service"\n fi\n done\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203888.94503: _low_level_execute_command(): starting 13731 1727203888.94506: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203888.668858-18395-167751050130994/ > /dev/null 2>&1 && sleep 0' 13731 1727203888.95174: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203888.95247: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203888.95282: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203888.95330: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203888.95390: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203888.97203: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203888.97218: stderr chunk (state=3): >>><<< 13731 1727203888.97228: stdout chunk (state=3): >>><<< 13731 1727203888.97482: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203888.97486: handler run complete 13731 1727203888.97488: Evaluated conditional (False): False 13731 1727203888.97490: attempt loop complete, returning result 13731 1727203888.97492: _execute() done 13731 1727203888.97494: dumping result to json 13731 1727203888.97496: done dumping result, returning 13731 1727203888.97498: done running TaskExecutor() for managed-node3/TASK: Stop dnsmasq/radvd services [028d2410-947f-82dc-c122-000000000e58] 13731 1727203888.97500: sending task result for task 028d2410-947f-82dc-c122-000000000e58 13731 1727203888.97572: done sending task result for task 028d2410-947f-82dc-c122-000000000e58 13731 1727203888.97577: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "set -uxo pipefail\nexec 1>&2\npkill -F /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.pid\nrm -rf /run/dhcp_testbr.lease\nif grep 'release 6' /etc/redhat-release; then\n # Stop radvd server\n service radvd stop\n iptables -D INPUT -i testbr -p udp --dport 67:68 --sport 67:68 -j ACCEPT\nfi\nif systemctl is-active firewalld; then\n for service in dhcp dhcpv6 dhcpv6-client; do\n if firewall-cmd --query-service=\"$service\"; then\n firewall-cmd --remove-service \"$service\"\n fi\n done\nfi\n", "delta": "0:00:00.025623", "end": "2024-09-24 14:51:28.928183", "rc": 0, "start": "2024-09-24 14:51:28.902560" } STDERR: + exec + pkill -F /run/dhcp_testbr.pid + rm -rf /run/dhcp_testbr.pid + rm -rf /run/dhcp_testbr.lease + grep 'release 6' /etc/redhat-release + systemctl is-active firewalld inactive 13731 1727203888.97647: no more pending results, returning what we have 13731 1727203888.97652: results queue empty 13731 1727203888.97652: checking for any_errors_fatal 13731 1727203888.97663: done checking for any_errors_fatal 13731 1727203888.97663: checking for max_fail_percentage 13731 1727203888.97665: done checking for max_fail_percentage 13731 1727203888.97666: checking to see if all hosts have failed and the running result is not ok 13731 1727203888.97667: done checking to see if all hosts have failed 13731 1727203888.97668: getting the remaining hosts for this loop 13731 1727203888.97670: done getting the remaining hosts for this loop 13731 1727203888.97673: getting the next task for host managed-node3 13731 1727203888.97688: done getting next task for host managed-node3 13731 1727203888.97692: ^ task is: TASK: Check routes and DNS 13731 1727203888.97696: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=4, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203888.97701: getting variables 13731 1727203888.97702: in VariableManager get_vars() 13731 1727203888.97750: Calling all_inventory to load vars for managed-node3 13731 1727203888.97754: Calling groups_inventory to load vars for managed-node3 13731 1727203888.97756: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203888.97773: Calling all_plugins_play to load vars for managed-node3 13731 1727203888.97785: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203888.97790: Calling groups_plugins_play to load vars for managed-node3 13731 1727203888.99573: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203889.01184: done with get_vars() 13731 1727203889.01208: done getting variables 13731 1727203889.01273: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Check routes and DNS] **************************************************** task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml:6 Tuesday 24 September 2024 14:51:29 -0400 (0:00:00.382) 0:01:07.244 ***** 13731 1727203889.01306: entering _queue_task() for managed-node3/shell 13731 1727203889.01659: worker is 1 (out of 1 available) 13731 1727203889.01672: exiting _queue_task() for managed-node3/shell 13731 1727203889.01797: done queuing things up, now waiting for results queue to drain 13731 1727203889.01800: waiting for pending results... 13731 1727203889.02090: running TaskExecutor() for managed-node3/TASK: Check routes and DNS 13731 1727203889.02126: in run() - task 028d2410-947f-82dc-c122-000000000e5c 13731 1727203889.02150: variable 'ansible_search_path' from source: unknown 13731 1727203889.02157: variable 'ansible_search_path' from source: unknown 13731 1727203889.02199: calling self._execute() 13731 1727203889.02339: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203889.02347: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203889.02350: variable 'omit' from source: magic vars 13731 1727203889.02720: variable 'ansible_distribution_major_version' from source: facts 13731 1727203889.02885: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203889.02889: variable 'omit' from source: magic vars 13731 1727203889.02891: variable 'omit' from source: magic vars 13731 1727203889.02894: variable 'omit' from source: magic vars 13731 1727203889.02927: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203889.02975: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203889.03005: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203889.03029: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203889.03054: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203889.03096: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203889.03280: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203889.03283: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203889.03286: Set connection var ansible_pipelining to False 13731 1727203889.03288: Set connection var ansible_shell_type to sh 13731 1727203889.03290: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203889.03292: Set connection var ansible_connection to ssh 13731 1727203889.03294: Set connection var ansible_shell_executable to /bin/sh 13731 1727203889.03296: Set connection var ansible_timeout to 10 13731 1727203889.03298: variable 'ansible_shell_executable' from source: unknown 13731 1727203889.03300: variable 'ansible_connection' from source: unknown 13731 1727203889.03302: variable 'ansible_module_compression' from source: unknown 13731 1727203889.03304: variable 'ansible_shell_type' from source: unknown 13731 1727203889.03306: variable 'ansible_shell_executable' from source: unknown 13731 1727203889.03315: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203889.03324: variable 'ansible_pipelining' from source: unknown 13731 1727203889.03331: variable 'ansible_timeout' from source: unknown 13731 1727203889.03340: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203889.03500: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203889.03518: variable 'omit' from source: magic vars 13731 1727203889.03533: starting attempt loop 13731 1727203889.03542: running the handler 13731 1727203889.03557: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203889.03587: _low_level_execute_command(): starting 13731 1727203889.03602: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203889.04333: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203889.04348: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203889.04389: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.04404: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203889.04497: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.04519: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.04541: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.04564: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.04641: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.06193: stdout chunk (state=3): >>>/root <<< 13731 1727203889.06353: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.06356: stdout chunk (state=3): >>><<< 13731 1727203889.06359: stderr chunk (state=3): >>><<< 13731 1727203889.06390: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203889.06502: _low_level_execute_command(): starting 13731 1727203889.06506: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901 `" && echo ansible-tmp-1727203889.0639837-18412-176527923645901="` echo /root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901 `" ) && sleep 0' 13731 1727203889.07247: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.07288: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.07363: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.09213: stdout chunk (state=3): >>>ansible-tmp-1727203889.0639837-18412-176527923645901=/root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901 <<< 13731 1727203889.09377: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.09381: stdout chunk (state=3): >>><<< 13731 1727203889.09383: stderr chunk (state=3): >>><<< 13731 1727203889.09403: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203889.0639837-18412-176527923645901=/root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203889.09440: variable 'ansible_module_compression' from source: unknown 13731 1727203889.09581: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203889.09586: variable 'ansible_facts' from source: unknown 13731 1727203889.09637: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/AnsiballZ_command.py 13731 1727203889.09831: Sending initial data 13731 1727203889.09840: Sent initial data (156 bytes) 13731 1727203889.10454: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203889.10563: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.10581: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.10623: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.10660: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.12153: stderr chunk (state=3): >>>debug2: Remote version: 3 <<< 13731 1727203889.12169: stderr chunk (state=3): >>>debug2: Server supports extension "posix-rename@openssh.com" revision 1 <<< 13731 1727203889.12186: stderr chunk (state=3): >>>debug2: Server supports extension "statvfs@openssh.com" revision 2 <<< 13731 1727203889.12219: stderr chunk (state=3): >>>debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203889.12266: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203889.12306: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpx91lgps5 /root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/AnsiballZ_command.py <<< 13731 1727203889.12322: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/AnsiballZ_command.py" <<< 13731 1727203889.12360: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmpx91lgps5" to remote "/root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/AnsiballZ_command.py" debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/AnsiballZ_command.py" <<< 13731 1727203889.13114: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.13308: stderr chunk (state=3): >>><<< 13731 1727203889.13311: stdout chunk (state=3): >>><<< 13731 1727203889.13313: done transferring module to remote 13731 1727203889.13315: _low_level_execute_command(): starting 13731 1727203889.13318: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/ /root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/AnsiballZ_command.py && sleep 0' 13731 1727203889.13921: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203889.13926: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.13963: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.13986: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.14009: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.14087: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.15840: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.15868: stdout chunk (state=3): >>><<< 13731 1727203889.15874: stderr chunk (state=3): >>><<< 13731 1727203889.15994: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203889.15998: _low_level_execute_command(): starting 13731 1727203889.16002: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/AnsiballZ_command.py && sleep 0' 13731 1727203889.16632: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.16700: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.16717: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.16745: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.16817: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.32486: stdout chunk (state=3): >>> {"changed": true, "stdout": "IP\n1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n inet 127.0.0.1/8 scope host lo\n valid_lft forever preferred_lft forever\n inet6 ::1/128 scope host noprefixroute \n valid_lft forever preferred_lft forever\n2: eth0: mtu 9001 qdisc mq state UP group default qlen 1000\n link/ether 02:83:38:1a:ae:4d brd ff:ff:ff:ff:ff:ff\n altname enX0\n inet 10.31.47.22/22 brd 10.31.47.255 scope global dynamic noprefixroute eth0\n valid_lft 3164sec preferred_lft 3164sec\n inet6 fe80::83:38ff:fe1a:ae4d/64 scope link noprefixroute \n valid_lft forever preferred_lft forever\nIP ROUTE\ndefault via 10.31.44.1 dev eth0 proto dhcp src 10.31.47.22 metric 100 \n10.31.44.0/22 dev eth0 proto kernel scope link src 10.31.47.22 metric 100 \nIP -6 ROUTE\nfe80::/64 dev eth0 proto kernel metric 1024 pref medium\nRESOLV\n# Generated by NetworkManager\nsearch us-east-1.aws.redhat.com\nnameserver 10.29.169.13\nnameserver 10.29.170.12\nnameserver 10.2.32.1", "stderr": "", "rc": 0, "cmd": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "start": "2024-09-24 14:51:29.315735", "end": "2024-09-24 14:51:29.324261", "delta": "0:00:00.008526", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203889.33849: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203889.33878: stderr chunk (state=3): >>><<< 13731 1727203889.33883: stdout chunk (state=3): >>><<< 13731 1727203889.33905: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "IP\n1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n inet 127.0.0.1/8 scope host lo\n valid_lft forever preferred_lft forever\n inet6 ::1/128 scope host noprefixroute \n valid_lft forever preferred_lft forever\n2: eth0: mtu 9001 qdisc mq state UP group default qlen 1000\n link/ether 02:83:38:1a:ae:4d brd ff:ff:ff:ff:ff:ff\n altname enX0\n inet 10.31.47.22/22 brd 10.31.47.255 scope global dynamic noprefixroute eth0\n valid_lft 3164sec preferred_lft 3164sec\n inet6 fe80::83:38ff:fe1a:ae4d/64 scope link noprefixroute \n valid_lft forever preferred_lft forever\nIP ROUTE\ndefault via 10.31.44.1 dev eth0 proto dhcp src 10.31.47.22 metric 100 \n10.31.44.0/22 dev eth0 proto kernel scope link src 10.31.47.22 metric 100 \nIP -6 ROUTE\nfe80::/64 dev eth0 proto kernel metric 1024 pref medium\nRESOLV\n# Generated by NetworkManager\nsearch us-east-1.aws.redhat.com\nnameserver 10.29.169.13\nnameserver 10.29.170.12\nnameserver 10.2.32.1", "stderr": "", "rc": 0, "cmd": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "start": "2024-09-24 14:51:29.315735", "end": "2024-09-24 14:51:29.324261", "delta": "0:00:00.008526", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203889.33946: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203889.33953: _low_level_execute_command(): starting 13731 1727203889.33958: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203889.0639837-18412-176527923645901/ > /dev/null 2>&1 && sleep 0' 13731 1727203889.34447: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203889.34450: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203889.34452: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.34455: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203889.34457: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203889.34459: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.34519: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.34522: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.34587: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.36349: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.36379: stderr chunk (state=3): >>><<< 13731 1727203889.36382: stdout chunk (state=3): >>><<< 13731 1727203889.36396: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203889.36403: handler run complete 13731 1727203889.36421: Evaluated conditional (False): False 13731 1727203889.36429: attempt loop complete, returning result 13731 1727203889.36433: _execute() done 13731 1727203889.36436: dumping result to json 13731 1727203889.36441: done dumping result, returning 13731 1727203889.36448: done running TaskExecutor() for managed-node3/TASK: Check routes and DNS [028d2410-947f-82dc-c122-000000000e5c] 13731 1727203889.36453: sending task result for task 028d2410-947f-82dc-c122-000000000e5c 13731 1727203889.36559: done sending task result for task 028d2410-947f-82dc-c122-000000000e5c 13731 1727203889.36564: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "delta": "0:00:00.008526", "end": "2024-09-24 14:51:29.324261", "rc": 0, "start": "2024-09-24 14:51:29.315735" } STDOUT: IP 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host noprefixroute valid_lft forever preferred_lft forever 2: eth0: mtu 9001 qdisc mq state UP group default qlen 1000 link/ether 02:83:38:1a:ae:4d brd ff:ff:ff:ff:ff:ff altname enX0 inet 10.31.47.22/22 brd 10.31.47.255 scope global dynamic noprefixroute eth0 valid_lft 3164sec preferred_lft 3164sec inet6 fe80::83:38ff:fe1a:ae4d/64 scope link noprefixroute valid_lft forever preferred_lft forever IP ROUTE default via 10.31.44.1 dev eth0 proto dhcp src 10.31.47.22 metric 100 10.31.44.0/22 dev eth0 proto kernel scope link src 10.31.47.22 metric 100 IP -6 ROUTE fe80::/64 dev eth0 proto kernel metric 1024 pref medium RESOLV # Generated by NetworkManager search us-east-1.aws.redhat.com nameserver 10.29.169.13 nameserver 10.29.170.12 nameserver 10.2.32.1 13731 1727203889.36634: no more pending results, returning what we have 13731 1727203889.36638: results queue empty 13731 1727203889.36639: checking for any_errors_fatal 13731 1727203889.36649: done checking for any_errors_fatal 13731 1727203889.36650: checking for max_fail_percentage 13731 1727203889.36652: done checking for max_fail_percentage 13731 1727203889.36655: checking to see if all hosts have failed and the running result is not ok 13731 1727203889.36656: done checking to see if all hosts have failed 13731 1727203889.36656: getting the remaining hosts for this loop 13731 1727203889.36658: done getting the remaining hosts for this loop 13731 1727203889.36664: getting the next task for host managed-node3 13731 1727203889.36672: done getting next task for host managed-node3 13731 1727203889.36675: ^ task is: TASK: Verify DNS and network connectivity 13731 1727203889.36680: ^ state is: HOST STATE: block=5, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=4, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203889.36690: getting variables 13731 1727203889.36691: in VariableManager get_vars() 13731 1727203889.36734: Calling all_inventory to load vars for managed-node3 13731 1727203889.36737: Calling groups_inventory to load vars for managed-node3 13731 1727203889.36739: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203889.36748: Calling all_plugins_play to load vars for managed-node3 13731 1727203889.36751: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203889.36753: Calling groups_plugins_play to load vars for managed-node3 13731 1727203889.37565: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203889.38908: done with get_vars() 13731 1727203889.38932: done getting variables 13731 1727203889.38985: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Verify DNS and network connectivity] ************************************* task path: /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml:24 Tuesday 24 September 2024 14:51:29 -0400 (0:00:00.377) 0:01:07.621 ***** 13731 1727203889.39011: entering _queue_task() for managed-node3/shell 13731 1727203889.39284: worker is 1 (out of 1 available) 13731 1727203889.39297: exiting _queue_task() for managed-node3/shell 13731 1727203889.39309: done queuing things up, now waiting for results queue to drain 13731 1727203889.39310: waiting for pending results... 13731 1727203889.39502: running TaskExecutor() for managed-node3/TASK: Verify DNS and network connectivity 13731 1727203889.39582: in run() - task 028d2410-947f-82dc-c122-000000000e5d 13731 1727203889.39595: variable 'ansible_search_path' from source: unknown 13731 1727203889.39599: variable 'ansible_search_path' from source: unknown 13731 1727203889.39631: calling self._execute() 13731 1727203889.39711: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203889.39715: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203889.39723: variable 'omit' from source: magic vars 13731 1727203889.40020: variable 'ansible_distribution_major_version' from source: facts 13731 1727203889.40030: Evaluated conditional (ansible_distribution_major_version != '6'): True 13731 1727203889.40129: variable 'ansible_facts' from source: unknown 13731 1727203889.40764: Evaluated conditional (ansible_facts["distribution"] == "CentOS"): True 13731 1727203889.40768: variable 'omit' from source: magic vars 13731 1727203889.40807: variable 'omit' from source: magic vars 13731 1727203889.40832: variable 'omit' from source: magic vars 13731 1727203889.40887: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 13731 1727203889.40913: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 13731 1727203889.40929: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 13731 1727203889.40942: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203889.40953: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 13731 1727203889.40984: variable 'inventory_hostname' from source: host vars for 'managed-node3' 13731 1727203889.40988: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203889.40990: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203889.41058: Set connection var ansible_pipelining to False 13731 1727203889.41067: Set connection var ansible_shell_type to sh 13731 1727203889.41073: Set connection var ansible_module_compression to ZIP_DEFLATED 13731 1727203889.41078: Set connection var ansible_connection to ssh 13731 1727203889.41085: Set connection var ansible_shell_executable to /bin/sh 13731 1727203889.41090: Set connection var ansible_timeout to 10 13731 1727203889.41107: variable 'ansible_shell_executable' from source: unknown 13731 1727203889.41110: variable 'ansible_connection' from source: unknown 13731 1727203889.41113: variable 'ansible_module_compression' from source: unknown 13731 1727203889.41115: variable 'ansible_shell_type' from source: unknown 13731 1727203889.41117: variable 'ansible_shell_executable' from source: unknown 13731 1727203889.41119: variable 'ansible_host' from source: host vars for 'managed-node3' 13731 1727203889.41122: variable 'ansible_pipelining' from source: unknown 13731 1727203889.41124: variable 'ansible_timeout' from source: unknown 13731 1727203889.41129: variable 'ansible_ssh_extra_args' from source: host vars for 'managed-node3' 13731 1727203889.41235: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203889.41244: variable 'omit' from source: magic vars 13731 1727203889.41248: starting attempt loop 13731 1727203889.41251: running the handler 13731 1727203889.41261: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 13731 1727203889.41285: _low_level_execute_command(): starting 13731 1727203889.41289: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 13731 1727203889.41805: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203889.41809: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.41812: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203889.41815: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.41872: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.41880: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.41883: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.41914: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.43478: stdout chunk (state=3): >>>/root <<< 13731 1727203889.43597: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.43601: stderr chunk (state=3): >>><<< 13731 1727203889.43604: stdout chunk (state=3): >>><<< 13731 1727203889.43626: _low_level_execute_command() done: rc=0, stdout=/root , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203889.43638: _low_level_execute_command(): starting 13731 1727203889.43648: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112 `" && echo ansible-tmp-1727203889.4362574-18430-256846826221112="` echo /root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112 `" ) && sleep 0' 13731 1727203889.44119: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203889.44123: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.44126: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration <<< 13731 1727203889.44128: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203889.44130: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.44179: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.44182: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.44185: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.44228: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.46067: stdout chunk (state=3): >>>ansible-tmp-1727203889.4362574-18430-256846826221112=/root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112 <<< 13731 1727203889.46174: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.46204: stderr chunk (state=3): >>><<< 13731 1727203889.46207: stdout chunk (state=3): >>><<< 13731 1727203889.46222: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1727203889.4362574-18430-256846826221112=/root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112 , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203889.46249: variable 'ansible_module_compression' from source: unknown 13731 1727203889.46299: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-13731wdkjbbyg/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 13731 1727203889.46328: variable 'ansible_facts' from source: unknown 13731 1727203889.46390: transferring module to remote /root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/AnsiballZ_command.py 13731 1727203889.46501: Sending initial data 13731 1727203889.46504: Sent initial data (156 bytes) 13731 1727203889.46948: stderr chunk (state=3): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203889.46951: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.46953: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203889.46955: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.47013: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.47021: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.47024: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.47051: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.48545: stderr chunk (state=3): >>>debug2: Remote version: 3 debug2: Server supports extension "posix-rename@openssh.com" revision 1 debug2: Server supports extension "statvfs@openssh.com" revision 2 debug2: Server supports extension "fstatvfs@openssh.com" revision 2 debug2: Server supports extension "hardlink@openssh.com" revision 1 debug2: Server supports extension "fsync@openssh.com" revision 1 debug2: Server supports extension "lsetstat@openssh.com" revision 1 debug2: Server supports extension "limits@openssh.com" revision 1 debug2: Server supports extension "expand-path@openssh.com" revision 1 debug2: Server supports extension "copy-data" revision 1 debug2: Unrecognised server extension "home-directory" debug2: Server supports extension "users-groups-by-id@openssh.com" revision 1 <<< 13731 1727203889.48583: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_REALPATH "." <<< 13731 1727203889.48616: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp2jqnjo4s /root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/AnsiballZ_command.py <<< 13731 1727203889.48619: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_STAT "/root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/AnsiballZ_command.py" <<< 13731 1727203889.48647: stderr chunk (state=3): >>>debug1: stat remote: No such file or directory debug2: sftp_upload: upload local "/root/.ansible/tmp/ansible-local-13731wdkjbbyg/tmp2jqnjo4s" to remote "/root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/AnsiballZ_command.py" <<< 13731 1727203889.48655: stderr chunk (state=3): >>>debug2: Sending SSH2_FXP_OPEN "/root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/AnsiballZ_command.py" <<< 13731 1727203889.49145: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.49192: stderr chunk (state=3): >>><<< 13731 1727203889.49195: stdout chunk (state=3): >>><<< 13731 1727203889.49238: done transferring module to remote 13731 1727203889.49248: _low_level_execute_command(): starting 13731 1727203889.49252: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/ /root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/AnsiballZ_command.py && sleep 0' 13731 1727203889.49709: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203889.49712: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found <<< 13731 1727203889.49719: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.49721: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203889.49724: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.49781: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.49793: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.49797: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.49816: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.51516: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.51545: stderr chunk (state=3): >>><<< 13731 1727203889.51548: stdout chunk (state=3): >>><<< 13731 1727203889.51565: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203889.51569: _low_level_execute_command(): starting 13731 1727203889.51571: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 /root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/AnsiballZ_command.py && sleep 0' 13731 1727203889.52014: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203889.52018: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.52037: stderr chunk (state=3): >>>debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.52096: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.52103: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.52105: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.52136: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.74469: stdout chunk (state=3): >>> {"changed": true, "stdout": "CHECK DNS AND CONNECTIVITY\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "stderr": " % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 305 100 305 0 0 16091 0 --:--:-- --:--:-- --:--:-- 16944\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 291 100 291 0 0 8572 0 --:--:-- --:--:-- --:--:-- 8818", "rc": 0, "cmd": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "start": "2024-09-24 14:51:29.670115", "end": "2024-09-24 14:51:29.744140", "delta": "0:00:00.074025", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 13731 1727203889.75999: stderr chunk (state=3): >>>debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. <<< 13731 1727203889.76023: stderr chunk (state=3): >>><<< 13731 1727203889.76026: stdout chunk (state=3): >>><<< 13731 1727203889.76046: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "CHECK DNS AND CONNECTIVITY\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "stderr": " % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 305 100 305 0 0 16091 0 --:--:-- --:--:-- --:--:-- 16944\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 291 100 291 0 0 8572 0 --:--:-- --:--:-- --:--:-- 8818", "rc": 0, "cmd": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "start": "2024-09-24 14:51:29.670115", "end": "2024-09-24 14:51:29.744140", "delta": "0:00:00.074025", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 Shared connection to 10.31.47.22 closed. 13731 1727203889.76083: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts "$host"; then\n echo FAILED to lookup host "$host"\n exit 1\n fi\n if ! curl -o /dev/null https://"$host"; then\n echo FAILED to contact host "$host"\n exit 1\n fi\ndone\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.17.4', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/', '_ansible_remote_tmp': '~/.ansible/tmp', '_ansible_ignore_unknown_opts': False, '_ansible_target_log_info': None}) 13731 1727203889.76118: _low_level_execute_command(): starting 13731 1727203889.76121: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1727203889.4362574-18430-256846826221112/ > /dev/null 2>&1 && sleep 0' 13731 1727203889.76724: stderr chunk (state=2): >>>OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 <<< 13731 1727203889.76730: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203889.76740: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203889.76754: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203889.76769: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203889.76778: stderr chunk (state=3): >>>debug2: match not found <<< 13731 1727203889.76788: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.76803: stderr chunk (state=3): >>>debug1: configuration requests final Match pass <<< 13731 1727203889.76811: stderr chunk (state=3): >>>debug2: resolve_canonicalize: hostname 10.31.47.22 is address <<< 13731 1727203889.76818: stderr chunk (state=3): >>>debug1: re-parsing configuration <<< 13731 1727203889.76830: stderr chunk (state=3): >>>debug1: Reading configuration data /root/.ssh/config <<< 13731 1727203889.76834: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config <<< 13731 1727203889.76847: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf <<< 13731 1727203889.76855: stderr chunk (state=3): >>>debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 <<< 13731 1727203889.76861: stderr chunk (state=3): >>>debug2: match found <<< 13731 1727203889.76875: stderr chunk (state=3): >>>debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config <<< 13731 1727203889.76944: stderr chunk (state=3): >>>debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' <<< 13731 1727203889.76956: stderr chunk (state=3): >>>debug2: fd 3 setting O_NONBLOCK <<< 13731 1727203889.76983: stderr chunk (state=3): >>>debug2: mux_client_hello_exchange: master version 4 <<< 13731 1727203889.77077: stderr chunk (state=3): >>>debug1: mux_client_request_session: master session id: 2 <<< 13731 1727203889.78837: stderr chunk (state=3): >>>debug2: Received exit status from master 0 <<< 13731 1727203889.78863: stderr chunk (state=3): >>><<< 13731 1727203889.78867: stdout chunk (state=3): >>><<< 13731 1727203889.78886: _low_level_execute_command() done: rc=0, stdout=, stderr=OpenSSH_9.8p1, OpenSSL 3.2.2 4 Jun 2024 debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match not found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: configuration requests final Match pass debug2: resolve_canonicalize: hostname 10.31.47.22 is address debug1: re-parsing configuration debug1: Reading configuration data /root/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Reading configuration data /etc/ssh/ssh_config.d/50-redhat.conf debug2: checking match for 'final all' host 10.31.47.22 originally 10.31.47.22 debug2: match found debug1: Reading configuration data /etc/crypto-policies/back-ends/openssh.config debug1: auto-mux: Trying existing master at '/root/.ansible/cp/2833a247f6' debug2: fd 3 setting O_NONBLOCK debug2: mux_client_hello_exchange: master version 4 debug1: mux_client_request_session: master session id: 2 debug2: Received exit status from master 0 13731 1727203889.78891: handler run complete 13731 1727203889.78908: Evaluated conditional (False): False 13731 1727203889.78916: attempt loop complete, returning result 13731 1727203889.78919: _execute() done 13731 1727203889.78921: dumping result to json 13731 1727203889.78927: done dumping result, returning 13731 1727203889.78934: done running TaskExecutor() for managed-node3/TASK: Verify DNS and network connectivity [028d2410-947f-82dc-c122-000000000e5d] 13731 1727203889.78938: sending task result for task 028d2410-947f-82dc-c122-000000000e5d 13731 1727203889.79041: done sending task result for task 028d2410-947f-82dc-c122-000000000e5d 13731 1727203889.79044: WORKER PROCESS EXITING ok: [managed-node3] => { "changed": false, "cmd": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "delta": "0:00:00.074025", "end": "2024-09-24 14:51:29.744140", "rc": 0, "start": "2024-09-24 14:51:29.670115" } STDOUT: CHECK DNS AND CONNECTIVITY 2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.fedoraproject.org 2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.fedoraproject.org 2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.fedoraproject.org 2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.fedoraproject.org 2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.fedoraproject.org 2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.fedoraproject.org 2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org 2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org STDERR: % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 305 100 305 0 0 16091 0 --:--:-- --:--:-- --:--:-- 16944 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 291 100 291 0 0 8572 0 --:--:-- --:--:-- --:--:-- 8818 13731 1727203889.79119: no more pending results, returning what we have 13731 1727203889.79124: results queue empty 13731 1727203889.79124: checking for any_errors_fatal 13731 1727203889.79133: done checking for any_errors_fatal 13731 1727203889.79134: checking for max_fail_percentage 13731 1727203889.79136: done checking for max_fail_percentage 13731 1727203889.79136: checking to see if all hosts have failed and the running result is not ok 13731 1727203889.79137: done checking to see if all hosts have failed 13731 1727203889.79138: getting the remaining hosts for this loop 13731 1727203889.79142: done getting the remaining hosts for this loop 13731 1727203889.79146: getting the next task for host managed-node3 13731 1727203889.79164: done getting next task for host managed-node3 13731 1727203889.79166: ^ task is: TASK: meta (flush_handlers) 13731 1727203889.79168: ^ state is: HOST STATE: block=6, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203889.79173: getting variables 13731 1727203889.79176: in VariableManager get_vars() 13731 1727203889.79222: Calling all_inventory to load vars for managed-node3 13731 1727203889.79225: Calling groups_inventory to load vars for managed-node3 13731 1727203889.79227: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203889.79237: Calling all_plugins_play to load vars for managed-node3 13731 1727203889.79239: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203889.79242: Calling groups_plugins_play to load vars for managed-node3 13731 1727203889.84321: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203889.85168: done with get_vars() 13731 1727203889.85188: done getting variables 13731 1727203889.85230: in VariableManager get_vars() 13731 1727203889.85241: Calling all_inventory to load vars for managed-node3 13731 1727203889.85242: Calling groups_inventory to load vars for managed-node3 13731 1727203889.85244: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203889.85247: Calling all_plugins_play to load vars for managed-node3 13731 1727203889.85249: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203889.85250: Calling groups_plugins_play to load vars for managed-node3 13731 1727203889.85872: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203889.86714: done with get_vars() 13731 1727203889.86731: done queuing things up, now waiting for results queue to drain 13731 1727203889.86733: results queue empty 13731 1727203889.86733: checking for any_errors_fatal 13731 1727203889.86736: done checking for any_errors_fatal 13731 1727203889.86736: checking for max_fail_percentage 13731 1727203889.86737: done checking for max_fail_percentage 13731 1727203889.86738: checking to see if all hosts have failed and the running result is not ok 13731 1727203889.86738: done checking to see if all hosts have failed 13731 1727203889.86738: getting the remaining hosts for this loop 13731 1727203889.86739: done getting the remaining hosts for this loop 13731 1727203889.86741: getting the next task for host managed-node3 13731 1727203889.86743: done getting next task for host managed-node3 13731 1727203889.86744: ^ task is: TASK: meta (flush_handlers) 13731 1727203889.86745: ^ state is: HOST STATE: block=7, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203889.86747: getting variables 13731 1727203889.86748: in VariableManager get_vars() 13731 1727203889.86757: Calling all_inventory to load vars for managed-node3 13731 1727203889.86759: Calling groups_inventory to load vars for managed-node3 13731 1727203889.86760: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203889.86764: Calling all_plugins_play to load vars for managed-node3 13731 1727203889.86766: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203889.86768: Calling groups_plugins_play to load vars for managed-node3 13731 1727203889.87460: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203889.88287: done with get_vars() 13731 1727203889.88300: done getting variables 13731 1727203889.88334: in VariableManager get_vars() 13731 1727203889.88344: Calling all_inventory to load vars for managed-node3 13731 1727203889.88346: Calling groups_inventory to load vars for managed-node3 13731 1727203889.88347: Calling all_plugins_inventory to load vars for managed-node3 13731 1727203889.88350: Calling all_plugins_play to load vars for managed-node3 13731 1727203889.88351: Calling groups_plugins_inventory to load vars for managed-node3 13731 1727203889.88353: Calling groups_plugins_play to load vars for managed-node3 13731 1727203889.88964: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 13731 1727203889.89891: done with get_vars() 13731 1727203889.89909: done queuing things up, now waiting for results queue to drain 13731 1727203889.89910: results queue empty 13731 1727203889.89911: checking for any_errors_fatal 13731 1727203889.89912: done checking for any_errors_fatal 13731 1727203889.89912: checking for max_fail_percentage 13731 1727203889.89913: done checking for max_fail_percentage 13731 1727203889.89913: checking to see if all hosts have failed and the running result is not ok 13731 1727203889.89914: done checking to see if all hosts have failed 13731 1727203889.89914: getting the remaining hosts for this loop 13731 1727203889.89915: done getting the remaining hosts for this loop 13731 1727203889.89917: getting the next task for host managed-node3 13731 1727203889.89919: done getting next task for host managed-node3 13731 1727203889.89919: ^ task is: None 13731 1727203889.89920: ^ state is: HOST STATE: block=8, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 13731 1727203889.89921: done queuing things up, now waiting for results queue to drain 13731 1727203889.89922: results queue empty 13731 1727203889.89922: checking for any_errors_fatal 13731 1727203889.89923: done checking for any_errors_fatal 13731 1727203889.89923: checking for max_fail_percentage 13731 1727203889.89923: done checking for max_fail_percentage 13731 1727203889.89924: checking to see if all hosts have failed and the running result is not ok 13731 1727203889.89924: done checking to see if all hosts have failed 13731 1727203889.89925: getting the next task for host managed-node3 13731 1727203889.89927: done getting next task for host managed-node3 13731 1727203889.89927: ^ task is: None 13731 1727203889.89928: ^ state is: HOST STATE: block=8, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False PLAY RECAP ********************************************************************* managed-node3 : ok=148 changed=4 unreachable=0 failed=0 skipped=97 rescued=0 ignored=0 Tuesday 24 September 2024 14:51:29 -0400 (0:00:00.509) 0:01:08.130 ***** =============================================================================== ** TEST check bond settings --------------------------------------------- 6.68s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml:3 ** TEST check IPv4 ------------------------------------------------------ 2.76s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_IPv4_present.yml:3 fedora.linux_system_roles.network : Check which services are running ---- 1.92s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 ** TEST check bond settings --------------------------------------------- 1.91s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_bond_options.yml:3 fedora.linux_system_roles.network : Check which services are running ---- 1.87s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 fedora.linux_system_roles.network : Check which services are running ---- 1.85s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 fedora.linux_system_roles.network : Check which services are running ---- 1.82s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:21 Gathering Facts --------------------------------------------------------- 1.80s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/tests_bond_options_nm.yml:6 Create test interfaces -------------------------------------------------- 1.76s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:35 Create test interfaces -------------------------------------------------- 1.67s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:35 fedora.linux_system_roles.network : Check which packages are installed --- 1.16s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 fedora.linux_system_roles.network : Configure networking connection profiles --- 1.12s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 fedora.linux_system_roles.network : Enable and start NetworkManager ----- 1.11s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:122 fedora.linux_system_roles.network : Configure networking connection profiles --- 1.09s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 fedora.linux_system_roles.network : Check which packages are installed --- 1.08s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 Gathering Facts --------------------------------------------------------- 1.07s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_bond_options.yml:3 fedora.linux_system_roles.network : Check which packages are installed --- 1.01s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 fedora.linux_system_roles.network : Check which packages are installed --- 0.94s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/set_facts.yml:26 fedora.linux_system_roles.network : Configure networking connection profiles --- 0.94s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/roles/network/tasks/main.yml:159 Install dnsmasq --------------------------------------------------------- 0.83s /tmp/collections-bGV/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_test_interfaces_with_dhcp.yml:3 13731 1727203889.90011: RUNNING CLEANUP